var/home/core/zuul-output/0000755000175000017500000000000015070175331014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070206746015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004626062215070206737017714 0ustar rootrootOct 04 10:50:21 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 10:50:21 crc restorecon[4576]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:21 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:50:22 crc restorecon[4576]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 10:50:23 crc kubenswrapper[4758]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:50:23 crc kubenswrapper[4758]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 10:50:23 crc kubenswrapper[4758]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:50:23 crc kubenswrapper[4758]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:50:23 crc kubenswrapper[4758]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 10:50:23 crc kubenswrapper[4758]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.105032 4758 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111542 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111572 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111580 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111587 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111594 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111601 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111607 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111615 4758 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111626 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111637 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111645 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111653 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111660 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111668 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111675 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111682 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111689 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111696 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111703 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111710 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111716 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111724 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111730 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111736 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111743 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111749 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111756 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111762 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111768 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111775 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111781 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111787 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111796 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111804 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111811 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111821 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111829 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111835 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111841 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111848 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111854 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111870 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111877 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111884 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.111968 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112596 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112607 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112613 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112620 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112626 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112633 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112641 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112647 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112655 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112661 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112667 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112675 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112682 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112704 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112711 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112724 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112731 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112738 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112761 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112773 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112779 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112790 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112808 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112825 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112838 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.112852 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114535 4758 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114627 4758 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114932 4758 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114966 4758 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114976 4758 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114982 4758 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.114991 4758 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115001 4758 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115007 4758 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115013 4758 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115019 4758 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115025 4758 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115031 4758 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115036 4758 flags.go:64] FLAG: --cgroup-root="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115041 4758 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115046 4758 flags.go:64] FLAG: --client-ca-file="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115052 4758 flags.go:64] FLAG: --cloud-config="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115057 4758 flags.go:64] FLAG: --cloud-provider="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115062 4758 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115069 4758 flags.go:64] FLAG: --cluster-domain="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115074 4758 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115079 4758 flags.go:64] FLAG: --config-dir="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115085 4758 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115090 4758 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115101 4758 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115119 4758 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115128 4758 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115134 4758 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115139 4758 flags.go:64] FLAG: --contention-profiling="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115144 4758 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115150 4758 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115158 4758 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115187 4758 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115219 4758 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115225 4758 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115233 4758 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115238 4758 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115243 4758 flags.go:64] FLAG: --enable-server="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115248 4758 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115259 4758 flags.go:64] FLAG: --event-burst="100" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115264 4758 flags.go:64] FLAG: --event-qps="50" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115269 4758 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115274 4758 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115279 4758 flags.go:64] FLAG: --eviction-hard="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115286 4758 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115291 4758 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115297 4758 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115303 4758 flags.go:64] FLAG: --eviction-soft="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115309 4758 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115315 4758 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115320 4758 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115325 4758 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115331 4758 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115336 4758 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115341 4758 flags.go:64] FLAG: --feature-gates="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115349 4758 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115354 4758 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115360 4758 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115366 4758 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115372 4758 flags.go:64] FLAG: --healthz-port="10248" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115377 4758 flags.go:64] FLAG: --help="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115382 4758 flags.go:64] FLAG: --hostname-override="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115388 4758 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115394 4758 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115399 4758 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115405 4758 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115411 4758 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115418 4758 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115423 4758 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115428 4758 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115435 4758 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115440 4758 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115446 4758 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115451 4758 flags.go:64] FLAG: --kube-reserved="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115456 4758 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115461 4758 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115467 4758 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115473 4758 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115479 4758 flags.go:64] FLAG: --lock-file="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115484 4758 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115489 4758 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115495 4758 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115505 4758 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115510 4758 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115515 4758 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115520 4758 flags.go:64] FLAG: --logging-format="text" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115525 4758 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115530 4758 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115535 4758 flags.go:64] FLAG: --manifest-url="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115540 4758 flags.go:64] FLAG: --manifest-url-header="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115550 4758 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115555 4758 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115561 4758 flags.go:64] FLAG: --max-pods="110" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115566 4758 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115572 4758 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115578 4758 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115583 4758 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115588 4758 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115595 4758 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115600 4758 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115619 4758 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115624 4758 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115629 4758 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115635 4758 flags.go:64] FLAG: --pod-cidr="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115640 4758 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115652 4758 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115657 4758 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115662 4758 flags.go:64] FLAG: --pods-per-core="0" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115668 4758 flags.go:64] FLAG: --port="10250" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115674 4758 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115679 4758 flags.go:64] FLAG: --provider-id="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115684 4758 flags.go:64] FLAG: --qos-reserved="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115689 4758 flags.go:64] FLAG: --read-only-port="10255" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115695 4758 flags.go:64] FLAG: --register-node="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115700 4758 flags.go:64] FLAG: --register-schedulable="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115705 4758 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115716 4758 flags.go:64] FLAG: --registry-burst="10" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115722 4758 flags.go:64] FLAG: --registry-qps="5" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115727 4758 flags.go:64] FLAG: --reserved-cpus="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115731 4758 flags.go:64] FLAG: --reserved-memory="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115738 4758 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115744 4758 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115749 4758 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115755 4758 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115760 4758 flags.go:64] FLAG: --runonce="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115765 4758 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115770 4758 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115776 4758 flags.go:64] FLAG: --seccomp-default="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115782 4758 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115787 4758 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115793 4758 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115800 4758 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115806 4758 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115811 4758 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115817 4758 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115822 4758 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115827 4758 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115832 4758 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115838 4758 flags.go:64] FLAG: --system-cgroups="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115843 4758 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115854 4758 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115860 4758 flags.go:64] FLAG: --tls-cert-file="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115865 4758 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115873 4758 flags.go:64] FLAG: --tls-min-version="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115879 4758 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115884 4758 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115890 4758 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115895 4758 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115900 4758 flags.go:64] FLAG: --v="2" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115908 4758 flags.go:64] FLAG: --version="false" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115916 4758 flags.go:64] FLAG: --vmodule="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115923 4758 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.115929 4758 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116118 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116126 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116131 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116159 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116167 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116172 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116176 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116182 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116187 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116192 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116207 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116212 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116217 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116221 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116226 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116230 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116235 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116240 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116245 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116250 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116255 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116260 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116264 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116270 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116274 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116279 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116283 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116287 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116291 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116296 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116301 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116305 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116310 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116314 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116319 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116324 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116329 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116333 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116337 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116341 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116347 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116351 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116356 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116360 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116365 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116371 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116383 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116388 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116392 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116396 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116400 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116404 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116408 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116413 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116417 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116422 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116429 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116434 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116439 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116445 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116452 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116457 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116462 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116468 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116473 4758 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116478 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116482 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116487 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116491 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116497 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.116510 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.116519 4758 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.127429 4758 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.127482 4758 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127547 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127556 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127560 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127564 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127568 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127573 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127578 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127582 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127586 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127590 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127594 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127598 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127601 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127605 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127609 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127616 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127622 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127626 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127630 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127634 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127658 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127662 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127666 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127670 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127674 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127678 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127682 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127686 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127693 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127698 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127703 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127708 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127712 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127718 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127724 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127730 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127735 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127740 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127744 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127749 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127753 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127758 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127762 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127767 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127772 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127776 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127780 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127785 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127791 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127796 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127801 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127806 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127812 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127817 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127822 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127826 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127830 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127834 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127839 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127844 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127849 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127853 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127857 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127862 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127866 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127870 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127874 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127878 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127882 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127888 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.127893 4758 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.127900 4758 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128040 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128045 4758 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128050 4758 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128055 4758 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128059 4758 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128063 4758 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128066 4758 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128070 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128073 4758 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128077 4758 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128081 4758 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128085 4758 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128090 4758 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128096 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128101 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128106 4758 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128111 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128115 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128123 4758 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128127 4758 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128131 4758 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128148 4758 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128153 4758 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128156 4758 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128160 4758 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128163 4758 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128167 4758 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128171 4758 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128174 4758 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128178 4758 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128182 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128185 4758 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128189 4758 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128192 4758 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128197 4758 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128200 4758 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128204 4758 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128207 4758 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128211 4758 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128216 4758 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128220 4758 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128225 4758 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128228 4758 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128232 4758 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128236 4758 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128240 4758 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128244 4758 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128248 4758 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128251 4758 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128255 4758 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128258 4758 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128262 4758 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128266 4758 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128269 4758 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128273 4758 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128276 4758 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128280 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128284 4758 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128287 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128291 4758 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128294 4758 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128298 4758 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128302 4758 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128305 4758 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128310 4758 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128314 4758 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128318 4758 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128322 4758 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128326 4758 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128330 4758 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.128336 4758 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.128341 4758 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.128558 4758 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.134300 4758 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.134436 4758 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.135849 4758 server.go:997] "Starting client certificate rotation" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.135891 4758 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.137595 4758 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-02 16:25:31.02590612 +0000 UTC Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.137820 4758 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1421h35m7.888094993s for next certificate rotation Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.169588 4758 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.172008 4758 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.187131 4758 log.go:25] "Validated CRI v1 runtime API" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.217454 4758 log.go:25] "Validated CRI v1 image API" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.220404 4758 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.226001 4758 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-10-44-45-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.226045 4758 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.240814 4758 manager.go:217] Machine: {Timestamp:2025-10-04 10:50:23.238189468 +0000 UTC m=+0.530840377 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1c23ca6e-c5ec-490c-92c3-929ef50a0040 BootID:17ee5017-c6ea-48c4-a1fa-99cd3efae497 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:12:14:7e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:12:14:7e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:82:4d:36 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:23:c8:34 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:dc:9e:62 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:fb:62:dd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:92:9b:c9:0a:41:bb Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:4e:04:f4:51:60:fa Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.241278 4758 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.241535 4758 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.243523 4758 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.243758 4758 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.243797 4758 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.244080 4758 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.244093 4758 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.244484 4758 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.244520 4758 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.244793 4758 state_mem.go:36] "Initialized new in-memory state store" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.244897 4758 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.247826 4758 kubelet.go:418] "Attempting to sync node with API server" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.247869 4758 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.247901 4758 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.247919 4758 kubelet.go:324] "Adding apiserver pod source" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.247935 4758 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.252263 4758 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.253327 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.253325 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.253454 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.253487 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.253728 4758 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.256339 4758 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257866 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257911 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257927 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257941 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257964 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257979 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.257996 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.258035 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.258055 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.258069 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.258105 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.258152 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.260294 4758 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.261622 4758 server.go:1280] "Started kubelet" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.262186 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.262665 4758 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.262666 4758 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.263412 4758 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 10:50:23 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.267883 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.267923 4758 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.269268 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 15:14:09.48403754 +0000 UTC Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.269351 4758 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2356h23m46.214692886s for next certificate rotation Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.270035 4758 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.270063 4758 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.272809 4758 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.273070 4758 factory.go:55] Registering systemd factory Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.273106 4758 factory.go:221] Registration of the systemd container factory successfully Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.275618 4758 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.275950 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="200ms" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.276264 4758 factory.go:153] Registering CRI-O factory Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.276289 4758 factory.go:221] Registration of the crio container factory successfully Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.276374 4758 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.276430 4758 factory.go:103] Registering Raw factory Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.276450 4758 manager.go:1196] Started watching for new ooms in manager Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.275485 4758 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b442939282ca4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 10:50:23.26159274 +0000 UTC m=+0.554243639,LastTimestamp:2025-10-04 10:50:23.26159274 +0000 UTC m=+0.554243639,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.277327 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.278129 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.278222 4758 manager.go:319] Starting recovery of all containers Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.279410 4758 server.go:460] "Adding debug handlers to kubelet server" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.284694 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.284983 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.285131 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.285446 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.285703 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.285864 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.285987 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.286113 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.286280 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.286463 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.286615 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.286742 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.286860 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287028 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287208 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287361 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287482 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287613 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287733 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287850 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.287977 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288094 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288260 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288384 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288506 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288645 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288786 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.288907 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289027 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289192 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289316 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289456 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289600 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289726 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.289892 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290034 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290227 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290353 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290531 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290663 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290781 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.290928 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291049 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291223 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291349 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291468 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291591 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291732 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291854 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.291973 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292105 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292258 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292421 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292550 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292673 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292797 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.292918 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293055 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293239 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293363 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293482 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293599 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293715 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.293858 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294042 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294269 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294394 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294512 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294654 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294779 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.294899 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295023 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295214 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295360 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295482 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295600 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295717 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295855 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.295983 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299730 4758 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299814 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299838 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299854 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299868 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299889 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299900 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299912 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299926 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299936 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299947 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.299958 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300836 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300855 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300868 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300880 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300892 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300904 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300916 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300952 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300967 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300979 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.300992 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301006 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301509 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301545 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301575 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301593 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301611 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301624 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301640 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301651 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301662 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301672 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301683 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301692 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301702 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301711 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301721 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301729 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301739 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301749 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301759 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301769 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301780 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301790 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301800 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301809 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301818 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301828 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301841 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301854 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301865 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301874 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301883 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301893 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.301902 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302686 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302713 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302729 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302743 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302755 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302765 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302774 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302783 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302794 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302803 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302812 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302822 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302831 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302840 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302849 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302862 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302871 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302881 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302891 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302899 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302909 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302919 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302928 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302936 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302945 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302954 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302963 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302972 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302980 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.302989 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.306739 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.306893 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307013 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307270 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307422 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307547 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307667 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307791 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.307929 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308059 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308216 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308342 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308458 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308592 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308713 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.308865 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.309097 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.309271 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.309426 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.310276 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.310418 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.310544 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.310765 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.310917 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311058 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311237 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311362 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311485 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311628 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311770 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.311893 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312014 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312175 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312309 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312451 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312580 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312699 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312818 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.312974 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.313109 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.313287 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.313427 4758 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.313547 4758 reconstruct.go:97] "Volume reconstruction finished" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.313653 4758 reconciler.go:26] "Reconciler: start to sync state" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.323094 4758 manager.go:324] Recovery completed Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.323170 4758 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.324467 4758 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.324504 4758 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.324524 4758 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.324561 4758 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.325150 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.325187 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.334947 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.336959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.337077 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.337196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.337892 4758 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.337910 4758 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.337930 4758 state_mem.go:36] "Initialized new in-memory state store" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.353540 4758 policy_none.go:49] "None policy: Start" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.354268 4758 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.354319 4758 state_mem.go:35] "Initializing new in-memory state store" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.376664 4758 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.417353 4758 manager.go:334] "Starting Device Plugin manager" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.417394 4758 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.417404 4758 server.go:79] "Starting device plugin registration server" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.417796 4758 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.417810 4758 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.418104 4758 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.418188 4758 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.418194 4758 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.424636 4758 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.424786 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.425159 4758 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.426002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.426039 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.426052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.426226 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.426328 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.426364 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427076 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427129 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427227 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427320 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427346 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427692 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427731 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.427882 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428016 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428043 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428636 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428665 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428678 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428743 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428825 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428903 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.428949 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.429700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.429782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.429728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.429834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.429794 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.429895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.430077 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.430120 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.430929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.430960 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.430971 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.476720 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="400ms" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516509 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516554 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516579 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516600 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516621 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516688 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516739 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516794 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516823 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516860 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516898 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516936 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516961 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.516989 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.518458 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.519436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.519464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.519476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.519498 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.519867 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619418 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619451 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619542 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619565 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619590 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619597 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619651 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619682 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619740 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619801 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619810 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619600 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619555 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619848 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619889 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619904 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619938 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619943 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619687 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619982 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619994 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.619911 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.620028 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.620049 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.620164 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.620217 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.721375 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.722882 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.722921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.722934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.722957 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.723493 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.762872 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.772291 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.789599 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.799066 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: I1004 10:50:23.804315 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.816720 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7d139952c96421867a26279646bb5cf51ac8bf09920dc5de284e0a853678325d WatchSource:0}: Error finding container 7d139952c96421867a26279646bb5cf51ac8bf09920dc5de284e0a853678325d: Status 404 returned error can't find the container with id 7d139952c96421867a26279646bb5cf51ac8bf09920dc5de284e0a853678325d Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.822551 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4752cac87b72b79e07f0250cc30e7d158d666f44b4471725388a55972c2f6e8a WatchSource:0}: Error finding container 4752cac87b72b79e07f0250cc30e7d158d666f44b4471725388a55972c2f6e8a: Status 404 returned error can't find the container with id 4752cac87b72b79e07f0250cc30e7d158d666f44b4471725388a55972c2f6e8a Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.828915 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-768644a0ae076f70fda8390e90dfbf5dff02e6e33539335fe14a8957357a68bf WatchSource:0}: Error finding container 768644a0ae076f70fda8390e90dfbf5dff02e6e33539335fe14a8957357a68bf: Status 404 returned error can't find the container with id 768644a0ae076f70fda8390e90dfbf5dff02e6e33539335fe14a8957357a68bf Oct 04 10:50:23 crc kubenswrapper[4758]: W1004 10:50:23.833766 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b903bf2576609e326df143b66d9dde83e4f43a99d211474e930eb9d8f24bf1d0 WatchSource:0}: Error finding container b903bf2576609e326df143b66d9dde83e4f43a99d211474e930eb9d8f24bf1d0: Status 404 returned error can't find the container with id b903bf2576609e326df143b66d9dde83e4f43a99d211474e930eb9d8f24bf1d0 Oct 04 10:50:23 crc kubenswrapper[4758]: E1004 10:50:23.878053 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="800ms" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.124246 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.125715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.125766 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.125774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.125799 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.126343 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Oct 04 10:50:24 crc kubenswrapper[4758]: W1004 10:50:24.160979 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.161046 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.263582 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.330882 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e05f76a3b7b5d0441cda14285975b0c343e6487bc5845938e3239f99e001b0cc"} Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.332747 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b903bf2576609e326df143b66d9dde83e4f43a99d211474e930eb9d8f24bf1d0"} Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.333952 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"768644a0ae076f70fda8390e90dfbf5dff02e6e33539335fe14a8957357a68bf"} Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.334759 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4752cac87b72b79e07f0250cc30e7d158d666f44b4471725388a55972c2f6e8a"} Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.335489 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d139952c96421867a26279646bb5cf51ac8bf09920dc5de284e0a853678325d"} Oct 04 10:50:24 crc kubenswrapper[4758]: W1004 10:50:24.349563 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.349670 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:24 crc kubenswrapper[4758]: W1004 10:50:24.438509 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.438585 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:24 crc kubenswrapper[4758]: W1004 10:50:24.503568 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.503678 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.678788 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="1.6s" Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.777549 4758 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b442939282ca4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 10:50:23.26159274 +0000 UTC m=+0.554243639,LastTimestamp:2025-10-04 10:50:23.26159274 +0000 UTC m=+0.554243639,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.926886 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.928598 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.928645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.928713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:24 crc kubenswrapper[4758]: I1004 10:50:24.928744 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:24 crc kubenswrapper[4758]: E1004 10:50:24.929385 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.263265 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.343348 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.343442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.343475 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.343494 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.343457 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.344968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.345024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.345042 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.345864 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747" exitCode=0 Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.346606 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.347336 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.351567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.351625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.351648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.354411 4758 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5f893355eb7a1279ee50f338f93fc28d4f58a1d5594753183d2e56ae096083d6" exitCode=0 Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.354528 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5f893355eb7a1279ee50f338f93fc28d4f58a1d5594753183d2e56ae096083d6"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.354583 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.354886 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.355582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.355617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.355629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.356604 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.356641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.356654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.357787 4758 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca" exitCode=0 Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.357843 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.357881 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.358812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.358852 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.358872 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.360706 4758 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834" exitCode=0 Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.360764 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834"} Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.360799 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.361736 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.361792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.361812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.846895 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:25 crc kubenswrapper[4758]: I1004 10:50:25.913346 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.263901 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:26 crc kubenswrapper[4758]: E1004 10:50:26.280331 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="3.2s" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.366466 4758 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a633aa196e2b9becc8277d5a9150c0f96f1b1253b4d72f27eb29a49c5b4473d6" exitCode=0 Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.366566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a633aa196e2b9becc8277d5a9150c0f96f1b1253b4d72f27eb29a49c5b4473d6"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.366668 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.368055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.368089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.368101 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.371368 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.371448 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.371474 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.371501 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.376749 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.376821 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.376842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.378389 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.378392 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.380275 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.380620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.380680 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.384389 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.384469 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.384537 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.384561 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4"} Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.386042 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.386182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.386267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:26 crc kubenswrapper[4758]: W1004 10:50:26.473970 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:26 crc kubenswrapper[4758]: E1004 10:50:26.474048 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.530216 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.531306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.531349 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.531364 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:26 crc kubenswrapper[4758]: I1004 10:50:26.531433 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:26 crc kubenswrapper[4758]: E1004 10:50:26.531848 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Oct 04 10:50:27 crc kubenswrapper[4758]: W1004 10:50:27.229502 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:27 crc kubenswrapper[4758]: E1004 10:50:27.229591 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.264088 4758 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:27 crc kubenswrapper[4758]: W1004 10:50:27.270467 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Oct 04 10:50:27 crc kubenswrapper[4758]: E1004 10:50:27.270668 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.391045 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba"} Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.391152 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.391164 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132"} Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.392554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.392611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.392630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395485 4758 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ce99d37599b0323c5c6f9539dc1b345423c13cd98c90203aba70d22250bb9cf6" exitCode=0 Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395516 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ce99d37599b0323c5c6f9539dc1b345423c13cd98c90203aba70d22250bb9cf6"} Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395589 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395617 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395658 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395627 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395721 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.395730 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.397699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.397750 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.397770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.397926 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.397998 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.398021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.399024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.399068 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.399085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.399296 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.399336 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:27 crc kubenswrapper[4758]: I1004 10:50:27.399353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.403581 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1563dbccc1d9d8e041e832d6625387d78c909f64ece5b32197db26cf4f5451a8"} Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.403652 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b477d9e462e8066ecc00e9d5aa1462b0a2d4e9f66afd827ea76972af274b248e"} Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.403677 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e51cbd7fac851ef703372b17758184a6c4e2be795f43889f65a7bc00dae4d087"} Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.403695 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"aa916ebb0d3a0e2fc46f4e2d4377516363466b4558b8241bf2ccc984c158d13c"} Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.403692 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.403781 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.405714 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.405764 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.405780 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.548332 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.570050 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.661201 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.661434 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.663057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.663126 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:28 crc kubenswrapper[4758]: I1004 10:50:28.663139 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.411565 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9c907e7009daf7bd1ff768d878b6ff0d8a3d8e07a78ac73a8600ec97bcaa1018"} Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.411652 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.411698 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.412925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.412968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.412987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.412930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.413020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.413038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.458145 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.458322 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.459326 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.459363 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.459374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.732128 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.734169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.734204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.734217 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:29 crc kubenswrapper[4758]: I1004 10:50:29.734247 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.414253 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.414394 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.415294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.415327 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.415344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.415532 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.415574 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.415584 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:30 crc kubenswrapper[4758]: I1004 10:50:30.585864 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 10:50:31 crc kubenswrapper[4758]: I1004 10:50:31.212663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 10:50:31 crc kubenswrapper[4758]: I1004 10:50:31.416966 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:31 crc kubenswrapper[4758]: I1004 10:50:31.417884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:31 crc kubenswrapper[4758]: I1004 10:50:31.417934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:31 crc kubenswrapper[4758]: I1004 10:50:31.417951 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.045260 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.045495 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.047249 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.047298 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.047314 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.418628 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.419493 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.419542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:32 crc kubenswrapper[4758]: I1004 10:50:32.419551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.056528 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.056708 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.057832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.057896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.057916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:33 crc kubenswrapper[4758]: E1004 10:50:33.425284 4758 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.574817 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.575128 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.580009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.580055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:33 crc kubenswrapper[4758]: I1004 10:50:33.580072 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:36 crc kubenswrapper[4758]: I1004 10:50:36.575504 4758 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 10:50:36 crc kubenswrapper[4758]: I1004 10:50:36.575603 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 10:50:37 crc kubenswrapper[4758]: W1004 10:50:37.485015 4758 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 04 10:50:37 crc kubenswrapper[4758]: I1004 10:50:37.485211 4758 trace.go:236] Trace[1907713109]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:50:27.483) (total time: 10001ms): Oct 04 10:50:37 crc kubenswrapper[4758]: Trace[1907713109]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:50:37.484) Oct 04 10:50:37 crc kubenswrapper[4758]: Trace[1907713109]: [10.001737417s] [10.001737417s] END Oct 04 10:50:37 crc kubenswrapper[4758]: E1004 10:50:37.485258 4758 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.191362 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.191414 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.197897 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.197952 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.436181 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.437676 4758 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba" exitCode=255 Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.437714 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba"} Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.437854 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.438469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.438490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.438497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.438896 4758 scope.go:117] "RemoveContainer" containerID="e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba" Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.579527 4758 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]log ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]etcd ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/priority-and-fairness-filter ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-apiextensions-informers ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-apiextensions-controllers ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/crd-informer-synced ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-system-namespaces-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 04 10:50:38 crc kubenswrapper[4758]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 04 10:50:38 crc kubenswrapper[4758]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/bootstrap-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/start-kube-aggregator-informers ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-registration-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-discovery-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]autoregister-completion ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-openapi-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 04 10:50:38 crc kubenswrapper[4758]: livez check failed Oct 04 10:50:38 crc kubenswrapper[4758]: I1004 10:50:38.579579 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.320713 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.442504 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.444197 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5"} Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.444365 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.448047 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.448208 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.448237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.464954 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.465236 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.466848 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.466904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:39 crc kubenswrapper[4758]: I1004 10:50:39.466923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.447133 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.447259 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.448357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.448406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.448422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.656243 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.656407 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.657240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.657268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.657276 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:40 crc kubenswrapper[4758]: I1004 10:50:40.668800 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.449797 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.449806 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.451341 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.451378 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.451407 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.451430 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.451431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:41 crc kubenswrapper[4758]: I1004 10:50:41.451575 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.190655 4758 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.194078 4758 trace.go:236] Trace[1829010777]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:50:30.540) (total time: 12653ms): Oct 04 10:50:43 crc kubenswrapper[4758]: Trace[1829010777]: ---"Objects listed" error: 12653ms (10:50:43.193) Oct 04 10:50:43 crc kubenswrapper[4758]: Trace[1829010777]: [12.65357934s] [12.65357934s] END Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.194134 4758 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.194428 4758 trace.go:236] Trace[1749349805]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:50:32.414) (total time: 10779ms): Oct 04 10:50:43 crc kubenswrapper[4758]: Trace[1749349805]: ---"Objects listed" error: 10779ms (10:50:43.194) Oct 04 10:50:43 crc kubenswrapper[4758]: Trace[1749349805]: [10.779529746s] [10.779529746s] END Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.194453 4758 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.197416 4758 trace.go:236] Trace[1407332277]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:50:31.591) (total time: 11605ms): Oct 04 10:50:43 crc kubenswrapper[4758]: Trace[1407332277]: ---"Objects listed" error: 11605ms (10:50:43.197) Oct 04 10:50:43 crc kubenswrapper[4758]: Trace[1407332277]: [11.605967857s] [11.605967857s] END Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.197443 4758 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.198788 4758 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.200765 4758 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.258184 4758 apiserver.go:52] "Watching apiserver" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.264060 4758 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.264351 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.264926 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.264998 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.265066 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.265367 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.265394 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.265448 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.265470 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.265528 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.265862 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270058 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270287 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270419 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270504 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270439 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270655 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.270755 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.273460 4758 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.275175 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.277584 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.301658 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.301970 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302019 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302044 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302164 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302199 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302218 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302236 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302252 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302290 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302344 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302360 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302377 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302395 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302413 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302430 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302448 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302465 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302487 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302512 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302529 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302546 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302565 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302629 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302667 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302685 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302701 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302717 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302733 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302749 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302763 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302795 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302811 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302826 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302843 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302858 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302873 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302889 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302906 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302908 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302960 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302979 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.302996 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303012 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303050 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303082 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303114 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303130 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303136 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303147 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303165 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303185 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303202 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303218 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303246 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303263 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303279 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303347 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303373 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303389 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303407 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303418 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303429 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303582 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303600 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303617 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303634 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303650 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303668 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303683 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303699 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303717 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303736 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303735 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303773 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303785 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303948 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.303963 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304070 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304115 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304132 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304139 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304173 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304173 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304195 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304215 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304233 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304234 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304249 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304287 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304304 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304329 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304353 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304373 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304416 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304432 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304453 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304471 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304490 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304512 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304530 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304548 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304583 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.317625 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304326 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304524 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304545 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.304623 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:50:43.804590871 +0000 UTC m=+21.097241760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320186 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320217 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320241 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320262 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320280 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320303 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320330 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320352 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320371 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320389 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320410 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320428 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320449 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320467 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320484 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320547 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320595 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320612 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320628 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320644 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320664 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320681 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320709 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320726 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320743 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320761 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320778 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320799 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320816 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320834 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320855 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320871 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320889 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320908 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320949 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320968 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320985 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321019 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321036 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321053 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321071 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321088 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321122 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321138 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321155 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321171 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321191 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321207 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321224 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321241 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321257 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321273 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321291 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321308 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321325 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321341 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321357 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321374 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321390 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321406 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321421 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321437 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321452 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321485 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321501 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321519 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321537 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321555 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321588 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321625 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321641 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321675 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321691 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321724 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321740 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321757 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321774 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321791 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321809 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321825 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321841 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321857 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321873 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321889 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321905 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321921 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321939 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321956 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321972 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322007 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322029 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322055 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322074 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322134 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322170 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322188 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322206 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322224 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322241 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322259 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322289 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322343 4758 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322355 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322366 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322376 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322386 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322396 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322405 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322415 4758 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322427 4758 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322438 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322448 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322457 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.323214 4758 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.330935 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343025 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320534 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320715 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304731 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304793 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304953 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305127 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305254 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305343 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305437 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305562 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305578 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305637 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305722 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305757 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.305865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.306914 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.307838 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.308391 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.308626 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.308811 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.308965 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.309264 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.309529 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.312368 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.312529 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.312698 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.313545 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.313747 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.314662 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.314820 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.315418 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.315603 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.315777 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.316246 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.316415 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.316664 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.316827 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.318334 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.318512 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.319384 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.319651 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.319830 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.320890 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321062 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.321532 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322702 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.322897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.323898 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.325501 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.325861 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.327169 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.327533 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.327795 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.327824 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.327955 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.329342 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.329710 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.330133 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.330584 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.330887 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.331401 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.331832 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.331707 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.331883 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.331930 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.332129 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.332342 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.304624 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.332552 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.332895 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.332989 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.333091 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.333241 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.333506 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.333961 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.333995 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.334462 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.335057 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.335217 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.335557 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.335814 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.336751 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.337181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.337386 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.337609 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.338067 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.338355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.338559 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.338558 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.338859 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.339065 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.339261 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.339610 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.339684 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.340156 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.340614 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.340768 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.341427 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.341957 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.342617 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.342820 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.342916 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343091 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343307 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343579 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.346651 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343751 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343828 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.343994 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.344081 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.345012 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.344991 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.345089 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.345295 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.345495 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.345522 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.346573 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:43.846556056 +0000 UTC m=+21.139206945 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.346946 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.346951 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.346990 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:43.846980956 +0000 UTC m=+21.139631845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.347351 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.347676 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.348637 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.348763 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.349689 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.350180 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.350517 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.350611 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.350690 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.351178 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.351662 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.352246 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.352417 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.360426 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.360560 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.360627 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.360633 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.360666 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361143 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361235 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361416 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361433 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361459 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361533 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361602 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361610 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361656 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361785 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361582 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.361848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.362154 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.362845 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.363833 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.376184 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.380383 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.381138 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.381969 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.381208 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.382474 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383174 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383193 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383207 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383268 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:43.883249698 +0000 UTC m=+21.175900587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383448 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383465 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383484 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.383511 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:43.883503824 +0000 UTC m=+21.176154713 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.384359 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.384660 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.387242 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.393351 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.393583 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.398325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.399068 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.417587 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.423455 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.424570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.424654 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.424762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.424885 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425003 4758 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425071 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425158 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425223 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425298 4758 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425364 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425419 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425502 4758 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425562 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425617 4758 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425676 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425789 4758 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425847 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425907 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425960 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426017 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426068 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426141 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426204 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426259 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426318 4758 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426383 4758 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426457 4758 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426526 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426595 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426654 4758 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426706 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426758 4758 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426810 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426860 4758 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426915 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.426967 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427017 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427067 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427141 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427218 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427271 4758 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427321 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427375 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427425 4758 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427475 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427534 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427593 4758 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427652 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427703 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427755 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427811 4758 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427867 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427923 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.427975 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428030 4758 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428087 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428195 4758 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428256 4758 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428346 4758 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428425 4758 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428497 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428557 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428636 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428708 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428782 4758 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428857 4758 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.428927 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.429035 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.425315 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.429133 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.429672 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.429743 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.429855 4758 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.429949 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.430044 4758 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.430159 4758 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.430248 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440197 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440525 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440598 4758 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440666 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440734 4758 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440797 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440895 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.440955 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441012 4758 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441067 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441138 4758 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441192 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441244 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441338 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441416 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441471 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441527 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441579 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441635 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441691 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441743 4758 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441797 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441853 4758 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441906 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.441960 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442016 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442066 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442136 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442201 4758 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442254 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442389 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442461 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442541 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442617 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442674 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442735 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442789 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442890 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.442963 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443017 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443077 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443210 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443273 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443324 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443388 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443445 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443503 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443561 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443647 4758 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443703 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443766 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443833 4758 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443892 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.443958 4758 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444021 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444073 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444141 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444204 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444260 4758 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444314 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444387 4758 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444450 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444506 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444560 4758 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444620 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444675 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444730 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444783 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444836 4758 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444897 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.444964 4758 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445028 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445084 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445212 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445270 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445330 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445391 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445447 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445499 4758 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445556 4758 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445611 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445678 4758 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445745 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445798 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445848 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445900 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.445961 4758 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.449170 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.449534 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.449612 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.449693 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.449896 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.449922 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.450247 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.454082 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.457582 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.458146 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.458660 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.458700 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.459155 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.459940 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.464535 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.467233 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.468048 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.471046 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.472598 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.486165 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.505561 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.515706 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.524497 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.534734 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546444 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546476 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546486 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546496 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546504 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546513 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546521 4758 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546529 4758 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546539 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546547 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546554 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546562 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546569 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546577 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546586 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.546593 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.577878 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.590062 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.596807 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.708339 4758 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.801829 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.802204 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.802226 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.802914 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.803538 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.805452 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.808021 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.811351 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.811809 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.813729 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.818564 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.824520 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.824938 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.825378 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.838357 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.849796 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.850000 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:50:44.849979125 +0000 UTC m=+22.142630024 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850435 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850470 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850539 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850551 4758 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850561 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850574 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850584 4758 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850593 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.850604 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.850698 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.850748 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:44.850735353 +0000 UTC m=+22.143386232 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.850791 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.850815 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:44.850806725 +0000 UTC m=+22.143457614 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.853383 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.860838 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.869002 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.877441 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.887736 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.896799 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.905973 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.915042 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.926870 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.934153 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.944923 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.946304 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.948516 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.951668 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.951866 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.951892 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952092 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952202 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952261 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952296 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952312 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952378 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:44.952358651 +0000 UTC m=+22.245009530 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: E1004 10:50:43.952537 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:44.952519875 +0000 UTC m=+22.245170764 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.955718 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.956020 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.969944 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:43 crc kubenswrapper[4758]: I1004 10:50:43.983887 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.007240 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.031549 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.048092 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.081414 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.090854 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.101879 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.118145 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.127737 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.145169 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.160534 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.175692 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.193186 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.464365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9"} Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.464408 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"70612bc313279d145cb7c1e74b623ffa024546b07b29b5db92b2644bf1c116fb"} Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.465630 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed"} Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.465672 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"df2afba75f3b649fff4ee530169183790935776bb22784fb3823c1c3cc6478ca"} Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.468695 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"815f5a4ed4f9bf34bd1d5343f64de2e70699d0bd443b7ab7697feab32196e8af"} Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.861223 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.861296 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.861316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.861409 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.861438 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.861496 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:50:46.861459683 +0000 UTC m=+24.154110612 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.861541 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:46.861526245 +0000 UTC m=+24.154177174 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.861599 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:46.861588166 +0000 UTC m=+24.154239095 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.962091 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:44 crc kubenswrapper[4758]: I1004 10:50:44.962179 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962279 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962295 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962306 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962340 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962385 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962403 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962353 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:46.962340364 +0000 UTC m=+24.254991253 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:44 crc kubenswrapper[4758]: E1004 10:50:44.962515 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:46.962474977 +0000 UTC m=+24.255125866 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.088494 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qkzfh"] Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.089011 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.092008 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.092614 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.095200 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.110077 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.127217 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.143384 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.164322 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.188038 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.209786 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-znxcj"] Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.210039 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.218139 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.218370 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.218392 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.218510 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.228490 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.247482 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.263946 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49zpc\" (UniqueName: \"kubernetes.io/projected/9d218843-c1d0-4118-a718-98a5503a09a7-kube-api-access-49zpc\") pod \"node-resolver-qkzfh\" (UID: \"9d218843-c1d0-4118-a718-98a5503a09a7\") " pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.263987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9d218843-c1d0-4118-a718-98a5503a09a7-hosts-file\") pod \"node-resolver-qkzfh\" (UID: \"9d218843-c1d0-4118-a718-98a5503a09a7\") " pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.300978 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.325006 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.325055 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:45 crc kubenswrapper[4758]: E1004 10:50:45.325141 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:45 crc kubenswrapper[4758]: E1004 10:50:45.325269 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.325448 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:45 crc kubenswrapper[4758]: E1004 10:50:45.325657 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.329563 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.330118 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.330899 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.332066 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.333033 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.333533 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.334169 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.334690 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.336979 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.337478 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.337951 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.338974 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.339523 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.339565 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.340326 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.340809 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.341650 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.342255 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.342607 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.343532 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.344088 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.344522 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.345436 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.345884 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.346905 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.347316 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.348456 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.349061 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.349913 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.350456 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.351806 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.352275 4758 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.352373 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.354823 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.355086 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.355778 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.356190 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.358227 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.360438 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.361625 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.362520 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.365412 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-serviceca\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.365463 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49zpc\" (UniqueName: \"kubernetes.io/projected/9d218843-c1d0-4118-a718-98a5503a09a7-kube-api-access-49zpc\") pod \"node-resolver-qkzfh\" (UID: \"9d218843-c1d0-4118-a718-98a5503a09a7\") " pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.365484 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-host\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.365505 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9d218843-c1d0-4118-a718-98a5503a09a7-hosts-file\") pod \"node-resolver-qkzfh\" (UID: \"9d218843-c1d0-4118-a718-98a5503a09a7\") " pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.365525 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qrpx\" (UniqueName: \"kubernetes.io/projected/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-kube-api-access-9qrpx\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.365885 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/9d218843-c1d0-4118-a718-98a5503a09a7-hosts-file\") pod \"node-resolver-qkzfh\" (UID: \"9d218843-c1d0-4118-a718-98a5503a09a7\") " pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.366325 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.366858 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.367679 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.368460 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.370880 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.371372 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.372342 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.372832 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.378584 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.379166 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.390040 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.402835 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49zpc\" (UniqueName: \"kubernetes.io/projected/9d218843-c1d0-4118-a718-98a5503a09a7-kube-api-access-49zpc\") pod \"node-resolver-qkzfh\" (UID: \"9d218843-c1d0-4118-a718-98a5503a09a7\") " pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.404169 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qkzfh" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.416332 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: W1004 10:50:45.421265 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d218843_c1d0_4118_a718_98a5503a09a7.slice/crio-78cd8f5b5ecf81ffea35e7a4bd2c663e13377875a5b498bb8733357fe157b0e7 WatchSource:0}: Error finding container 78cd8f5b5ecf81ffea35e7a4bd2c663e13377875a5b498bb8733357fe157b0e7: Status 404 returned error can't find the container with id 78cd8f5b5ecf81ffea35e7a4bd2c663e13377875a5b498bb8733357fe157b0e7 Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.438113 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.452751 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.466065 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-host\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.466127 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qrpx\" (UniqueName: \"kubernetes.io/projected/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-kube-api-access-9qrpx\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.466178 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-serviceca\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.466216 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-host\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.467198 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-serviceca\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.469766 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.471595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qkzfh" event={"ID":"9d218843-c1d0-4118-a718-98a5503a09a7","Type":"ContainerStarted","Data":"78cd8f5b5ecf81ffea35e7a4bd2c663e13377875a5b498bb8733357fe157b0e7"} Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.474581 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074"} Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.484601 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qrpx\" (UniqueName: \"kubernetes.io/projected/e8a97aad-9d66-4f7e-ad8a-d829942dd4c7-kube-api-access-9qrpx\") pod \"node-ca-znxcj\" (UID: \"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\") " pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.485615 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.505920 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.516722 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.520399 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-znxcj" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.526996 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.537824 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.548551 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.560475 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: W1004 10:50:45.567415 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a97aad_9d66_4f7e_ad8a_d829942dd4c7.slice/crio-0c73998c8dbdcb5b8d72171ade9ebff8cb0bcaa7676d0cfde2a46a026d35422b WatchSource:0}: Error finding container 0c73998c8dbdcb5b8d72171ade9ebff8cb0bcaa7676d0cfde2a46a026d35422b: Status 404 returned error can't find the container with id 0c73998c8dbdcb5b8d72171ade9ebff8cb0bcaa7676d0cfde2a46a026d35422b Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.572796 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.584798 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.592814 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-d4w2q"] Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.593340 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.595434 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.595577 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.595622 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.597628 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.597842 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.604012 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.615687 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.638631 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.653265 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.665038 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.677375 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.686351 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.702085 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.717846 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.731430 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.746291 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.757975 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.767873 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4fd2c8d-474b-48ac-a369-ce63fbf61413-proxy-tls\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.767925 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b4fd2c8d-474b-48ac-a369-ce63fbf61413-rootfs\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.767944 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4fd2c8d-474b-48ac-a369-ce63fbf61413-mcd-auth-proxy-config\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.768039 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zndv9\" (UniqueName: \"kubernetes.io/projected/b4fd2c8d-474b-48ac-a369-ce63fbf61413-kube-api-access-zndv9\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.768486 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.792576 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.816796 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.835490 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.868788 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4fd2c8d-474b-48ac-a369-ce63fbf61413-proxy-tls\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.868828 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b4fd2c8d-474b-48ac-a369-ce63fbf61413-rootfs\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.868854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4fd2c8d-474b-48ac-a369-ce63fbf61413-mcd-auth-proxy-config\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.868877 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zndv9\" (UniqueName: \"kubernetes.io/projected/b4fd2c8d-474b-48ac-a369-ce63fbf61413-kube-api-access-zndv9\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.869194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b4fd2c8d-474b-48ac-a369-ce63fbf61413-rootfs\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.869889 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4fd2c8d-474b-48ac-a369-ce63fbf61413-mcd-auth-proxy-config\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.872782 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4fd2c8d-474b-48ac-a369-ce63fbf61413-proxy-tls\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.887851 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zndv9\" (UniqueName: \"kubernetes.io/projected/b4fd2c8d-474b-48ac-a369-ce63fbf61413-kube-api-access-zndv9\") pod \"machine-config-daemon-d4w2q\" (UID: \"b4fd2c8d-474b-48ac-a369-ce63fbf61413\") " pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.902996 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:50:45 crc kubenswrapper[4758]: W1004 10:50:45.914014 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4fd2c8d_474b_48ac_a369_ce63fbf61413.slice/crio-e80669f1b9d494234f955d5a42379fc31405779e562d159933482c10d3512c4d WatchSource:0}: Error finding container e80669f1b9d494234f955d5a42379fc31405779e562d159933482c10d3512c4d: Status 404 returned error can't find the container with id e80669f1b9d494234f955d5a42379fc31405779e562d159933482c10d3512c4d Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.979865 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-l4q8s"] Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.980463 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.982285 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-bpvxr"] Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.982475 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bpvxr" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.982877 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.983000 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.983177 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.983272 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.984256 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.984366 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.984470 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 10:50:45 crc kubenswrapper[4758]: I1004 10:50:45.991979 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.003689 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.011921 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.021538 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.032788 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.042640 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.053433 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.063576 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070376 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-os-release\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070413 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-netns\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070436 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-system-cni-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070450 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-cnibin\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070467 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-system-cni-dir\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070482 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrzz\" (UniqueName: \"kubernetes.io/projected/af0a535d-83d3-4669-9245-fa99ddcb6a7d-kube-api-access-slrzz\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070502 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-os-release\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070555 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-kubelet\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070620 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070638 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-socket-dir-parent\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070656 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-hostroot\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070684 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-cni-multus\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070697 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-daemon-config\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070715 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-etc-kubernetes\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070747 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-cni-bin\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070781 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cni-binary-copy\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070813 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-multus-certs\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070831 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-cni-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070853 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cnibin\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070872 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3875fe1d-de8c-4266-bf2b-e07c633b85dc-cni-binary-copy\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070895 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070916 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-k8s-cni-cncf-io\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070930 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-conf-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.070945 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctvk5\" (UniqueName: \"kubernetes.io/projected/3875fe1d-de8c-4266-bf2b-e07c633b85dc-kube-api-access-ctvk5\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.073649 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.084411 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.098298 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.109016 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.121751 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.135959 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.159342 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.171980 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-os-release\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172016 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrzz\" (UniqueName: \"kubernetes.io/projected/af0a535d-83d3-4669-9245-fa99ddcb6a7d-kube-api-access-slrzz\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172043 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-kubelet\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172060 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172075 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-socket-dir-parent\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172091 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-hostroot\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172124 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-cni-multus\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-daemon-config\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172157 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-cni-bin\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-etc-kubernetes\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172167 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-kubelet\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172185 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cni-binary-copy\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172242 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-multus-certs\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172271 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-cni-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cnibin\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172320 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3875fe1d-de8c-4266-bf2b-e07c633b85dc-cni-binary-copy\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172336 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172353 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-k8s-cni-cncf-io\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172367 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-conf-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172384 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctvk5\" (UniqueName: \"kubernetes.io/projected/3875fe1d-de8c-4266-bf2b-e07c633b85dc-kube-api-access-ctvk5\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172383 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-os-release\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172410 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-os-release\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-netns\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172462 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-os-release\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-multus-certs\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172496 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-system-cni-dir\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172519 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-system-cni-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172529 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-cni-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172534 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-cnibin\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172551 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cnibin\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-k8s-cni-cncf-io\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172607 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-conf-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172751 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cni-binary-copy\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172828 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-run-netns\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-cni-multus\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172852 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-etc-kubernetes\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172868 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-host-var-lib-cni-bin\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172903 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-cnibin\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.172976 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/af0a535d-83d3-4669-9245-fa99ddcb6a7d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173015 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-system-cni-dir\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173018 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-hostroot\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173048 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-system-cni-dir\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173061 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-socket-dir-parent\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173193 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/af0a535d-83d3-4669-9245-fa99ddcb6a7d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173535 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3875fe1d-de8c-4266-bf2b-e07c633b85dc-cni-binary-copy\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.173567 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3875fe1d-de8c-4266-bf2b-e07c633b85dc-multus-daemon-config\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.184680 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.191679 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctvk5\" (UniqueName: \"kubernetes.io/projected/3875fe1d-de8c-4266-bf2b-e07c633b85dc-kube-api-access-ctvk5\") pod \"multus-bpvxr\" (UID: \"3875fe1d-de8c-4266-bf2b-e07c633b85dc\") " pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.191796 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrzz\" (UniqueName: \"kubernetes.io/projected/af0a535d-83d3-4669-9245-fa99ddcb6a7d-kube-api-access-slrzz\") pod \"multus-additional-cni-plugins-l4q8s\" (UID: \"af0a535d-83d3-4669-9245-fa99ddcb6a7d\") " pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.211928 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.222880 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.233210 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.244111 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.254475 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.264726 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.300083 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.319230 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.324998 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bpvxr" Oct 04 10:50:46 crc kubenswrapper[4758]: W1004 10:50:46.337350 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3875fe1d_de8c_4266_bf2b_e07c633b85dc.slice/crio-31bca2c81ee646b5bea2b1d480bd06e07cc82ad7eb81c2ff5e3acbdd7a729e6a WatchSource:0}: Error finding container 31bca2c81ee646b5bea2b1d480bd06e07cc82ad7eb81c2ff5e3acbdd7a729e6a: Status 404 returned error can't find the container with id 31bca2c81ee646b5bea2b1d480bd06e07cc82ad7eb81c2ff5e3acbdd7a729e6a Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.343487 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.348358 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hbnh9"] Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.349132 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.373842 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.392591 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.414122 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.432078 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.451275 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.472663 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474170 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-systemd-units\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474207 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-bin\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474223 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474240 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-script-lib\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474258 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-env-overrides\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474273 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-netns\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474287 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-ovn\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474311 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-config\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474327 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474341 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-node-log\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474354 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-log-socket\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474370 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063cec2b-782d-4033-a24c-4129a935cf27-ovn-node-metrics-cert\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474392 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-ovn-kubernetes\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474407 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-netd\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474421 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-slash\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474435 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-var-lib-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474449 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-systemd\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474462 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tk2m\" (UniqueName: \"kubernetes.io/projected/063cec2b-782d-4033-a24c-4129a935cf27-kube-api-access-5tk2m\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474484 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-kubelet\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.474500 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-etc-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.480248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.481962 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qkzfh" event={"ID":"9d218843-c1d0-4118-a718-98a5503a09a7","Type":"ContainerStarted","Data":"88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.482794 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerStarted","Data":"63ed9564267fcbfe3e74d99843c0540c51bd1fa68d03f3ca031cf2d737ce850f"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.483524 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-znxcj" event={"ID":"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7","Type":"ContainerStarted","Data":"ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.483546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-znxcj" event={"ID":"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7","Type":"ContainerStarted","Data":"0c73998c8dbdcb5b8d72171ade9ebff8cb0bcaa7676d0cfde2a46a026d35422b"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.484341 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerStarted","Data":"31bca2c81ee646b5bea2b1d480bd06e07cc82ad7eb81c2ff5e3acbdd7a729e6a"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.485418 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.485440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"e80669f1b9d494234f955d5a42379fc31405779e562d159933482c10d3512c4d"} Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.492285 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.520122 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.559817 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.575883 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-netns\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.575921 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-ovn\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.575951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-config\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.575967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.575985 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-node-log\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.575978 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-netns\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576001 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-log-socket\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576046 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-log-socket\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576062 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-ovn\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576081 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576067 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063cec2b-782d-4033-a24c-4129a935cf27-ovn-node-metrics-cert\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576139 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-node-log\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576189 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-ovn-kubernetes\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576208 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-netd\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-slash\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576242 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-var-lib-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576269 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-systemd\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576282 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tk2m\" (UniqueName: \"kubernetes.io/projected/063cec2b-782d-4033-a24c-4129a935cf27-kube-api-access-5tk2m\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576298 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-kubelet\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576312 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-etc-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-systemd-units\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576389 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-bin\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576448 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576473 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-script-lib\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576489 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-env-overrides\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576714 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-kubelet\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576721 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-config\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576792 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-systemd-units\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576819 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-etc-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.576998 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-env-overrides\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577073 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-bin\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577187 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-slash\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577201 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577235 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-var-lib-openvswitch\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577380 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-systemd\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577416 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-ovn-kubernetes\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577484 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-netd\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.577640 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-script-lib\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.581394 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063cec2b-782d-4033-a24c-4129a935cf27-ovn-node-metrics-cert\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.599990 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.627779 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tk2m\" (UniqueName: \"kubernetes.io/projected/063cec2b-782d-4033-a24c-4129a935cf27-kube-api-access-5tk2m\") pod \"ovnkube-node-hbnh9\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.660378 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.674031 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:46 crc kubenswrapper[4758]: W1004 10:50:46.683016 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod063cec2b_782d_4033_a24c_4129a935cf27.slice/crio-1d3131a1ae8707e55ac8d5d638f1d2e8de3048e55c1c148906fdf96d320dc1a7 WatchSource:0}: Error finding container 1d3131a1ae8707e55ac8d5d638f1d2e8de3048e55c1c148906fdf96d320dc1a7: Status 404 returned error can't find the container with id 1d3131a1ae8707e55ac8d5d638f1d2e8de3048e55c1c148906fdf96d320dc1a7 Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.703899 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.744071 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.783116 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.820352 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.861061 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.879019 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.879129 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.879156 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.879237 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:50:50.879210527 +0000 UTC m=+28.171861416 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.879240 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.879264 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.879303 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:50.879292339 +0000 UTC m=+28.171943228 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.879316 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:50.879310049 +0000 UTC m=+28.171960938 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.898358 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.943842 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.979728 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.979773 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.979900 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.979916 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.979936 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.979973 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.980013 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.980033 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.979985 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:50.979971094 +0000 UTC m=+28.272621983 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:46 crc kubenswrapper[4758]: E1004 10:50:46.980151 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:50.980125787 +0000 UTC m=+28.272776716 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:46 crc kubenswrapper[4758]: I1004 10:50:46.994254 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.021135 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.060635 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.102674 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.325678 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.325700 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:47 crc kubenswrapper[4758]: E1004 10:50:47.325819 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:47 crc kubenswrapper[4758]: E1004 10:50:47.325988 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.326248 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:47 crc kubenswrapper[4758]: E1004 10:50:47.326375 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.490064 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerStarted","Data":"13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1"} Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.492078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e"} Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.494507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerStarted","Data":"6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea"} Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.496044 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" exitCode=0 Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.496473 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.496504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"1d3131a1ae8707e55ac8d5d638f1d2e8de3048e55c1c148906fdf96d320dc1a7"} Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.506814 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.547232 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.560350 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.574193 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.600186 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.612252 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.641462 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.667060 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.684033 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.703310 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.723213 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.734796 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.751835 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.763131 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.781175 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.811210 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.822745 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.834508 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.859295 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.898265 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.963695 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:47 crc kubenswrapper[4758]: I1004 10:50:47.978813 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:47Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.024251 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.062296 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.100973 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.141455 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.183037 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.219793 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.503092 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.503173 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.503185 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.503194 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.503226 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.503237 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.504420 4758 generic.go:334] "Generic (PLEG): container finished" podID="af0a535d-83d3-4669-9245-fa99ddcb6a7d" containerID="6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea" exitCode=0 Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.504449 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerDied","Data":"6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea"} Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.530800 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.546981 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.552813 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.565235 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.576723 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.586662 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.599174 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.612609 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.625929 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.639548 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.660289 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.674814 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.698972 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.748665 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.785197 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.819354 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.860415 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.903754 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.940370 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:48 crc kubenswrapper[4758]: I1004 10:50:48.979294 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:48Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.020486 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.061594 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.102403 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.145024 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.183899 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.223983 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.258226 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.301637 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.326581 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.326678 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.326930 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.326960 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.327062 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.327198 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.386153 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.509149 4758 generic.go:334] "Generic (PLEG): container finished" podID="af0a535d-83d3-4669-9245-fa99ddcb6a7d" containerID="ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0" exitCode=0 Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.509192 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerDied","Data":"ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0"} Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.526638 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.540559 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.557312 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.572900 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.583286 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.597554 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.598907 4758 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.600796 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.600821 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.600832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.600926 4758 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.620147 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.673135 4758 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.673491 4758 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.674633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.674731 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.674750 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.674775 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.674824 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.695885 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.699587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.699624 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.699636 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.699653 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.699666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.705805 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.714804 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.717426 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.717456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.717465 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.717479 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.717489 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.729893 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.733235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.733262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.733270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.733282 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.733291 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.744349 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.747191 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.753411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.753476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.753497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.753533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.753551 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.773421 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: E1004 10:50:49.773540 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.775003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.775024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.775032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.775045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.775054 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.778869 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.820710 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.862050 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.877062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.877096 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.877122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.877138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.877148 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.901148 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.940753 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.979738 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.979784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.979794 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.979809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:49 crc kubenswrapper[4758]: I1004 10:50:49.979822 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:49Z","lastTransitionTime":"2025-10-04T10:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.082326 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.082385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.082405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.082430 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.082450 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.184942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.184971 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.184981 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.184994 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.185002 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.287287 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.287355 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.287373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.287396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.287417 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.389537 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.389570 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.389580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.389598 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.389607 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.491434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.491473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.491482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.491494 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.491504 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.550072 4758 generic.go:334] "Generic (PLEG): container finished" podID="af0a535d-83d3-4669-9245-fa99ddcb6a7d" containerID="01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f" exitCode=0 Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.550134 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerDied","Data":"01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.565929 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.580164 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.593990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.594013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.594021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.594034 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.594043 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.601813 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.622200 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.637564 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.649927 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.662175 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.671019 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.680781 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.691348 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.695881 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.695911 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.695921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.695934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.695944 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.702633 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.717892 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.732165 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.748379 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:50Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.798569 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.798604 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.798616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.798647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.798657 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.900762 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.900798 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.900810 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.900825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.900839 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:50Z","lastTransitionTime":"2025-10-04T10:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.949168 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.949237 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:50 crc kubenswrapper[4758]: I1004 10:50:50.949269 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:50 crc kubenswrapper[4758]: E1004 10:50:50.949380 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:50 crc kubenswrapper[4758]: E1004 10:50:50.949412 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:50 crc kubenswrapper[4758]: E1004 10:50:50.949428 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:58.949413807 +0000 UTC m=+36.242064696 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:50 crc kubenswrapper[4758]: E1004 10:50:50.949511 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:58.949488069 +0000 UTC m=+36.242138968 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:50 crc kubenswrapper[4758]: E1004 10:50:50.949579 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:50:58.949568661 +0000 UTC m=+36.242219560 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.002927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.002962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.002971 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.002985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.002994 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.050094 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.050245 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050410 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050436 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050457 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050709 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:59.050682917 +0000 UTC m=+36.343333846 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050875 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050906 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050917 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.050970 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:50:59.050948203 +0000 UTC m=+36.343599182 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.105542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.105581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.105592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.105607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.105616 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.207626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.207667 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.207676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.207690 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.207702 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.310379 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.310618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.310698 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.310756 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.310810 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.326323 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.326496 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.326355 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.326718 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.326839 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:51 crc kubenswrapper[4758]: E1004 10:50:51.326946 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.413123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.413374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.413446 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.413504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.413557 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.515250 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.515359 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.515441 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.515527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.515611 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.556808 4758 generic.go:334] "Generic (PLEG): container finished" podID="af0a535d-83d3-4669-9245-fa99ddcb6a7d" containerID="741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b" exitCode=0 Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.557007 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerDied","Data":"741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.570514 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.573718 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.585745 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.605971 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.618581 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.618832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.618856 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.618866 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.618884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.618895 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.664313 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.678803 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.688172 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.699940 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.711943 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.723445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.723491 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.723505 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.723524 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.723537 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.723838 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.741119 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.754707 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.770276 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.781738 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:51Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.826485 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.826506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.826514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.826527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.826536 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.929630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.929669 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.929677 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.929692 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:51 crc kubenswrapper[4758]: I1004 10:50:51.929701 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:51Z","lastTransitionTime":"2025-10-04T10:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.033113 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.033141 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.033150 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.033163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.033173 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.136005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.136037 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.136048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.136064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.136077 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.239569 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.239625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.239644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.239665 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.239681 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.343051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.343154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.343176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.343201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.343219 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.446270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.446314 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.446326 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.446344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.446356 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.549329 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.549390 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.549409 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.549431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.549449 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.578740 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerStarted","Data":"519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.600990 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.623172 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.643616 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.653311 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.653397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.653420 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.653453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.653476 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.664661 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.677696 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.694812 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.723216 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.739069 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.756518 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.756576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.756592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.756616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.756633 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.758649 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.780241 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.800964 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.817296 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.834807 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.852644 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.859002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.859052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.859068 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.859161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.859181 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.961791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.961844 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.961863 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.961890 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:52 crc kubenswrapper[4758]: I1004 10:50:52.961907 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:52Z","lastTransitionTime":"2025-10-04T10:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.064024 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.064089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.064154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.064183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.064207 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.168172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.168234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.168250 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.168274 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.168291 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.272199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.272857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.273187 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.274011 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.274247 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.325509 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:53 crc kubenswrapper[4758]: E1004 10:50:53.325672 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.325958 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.326237 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:53 crc kubenswrapper[4758]: E1004 10:50:53.326351 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:53 crc kubenswrapper[4758]: E1004 10:50:53.326241 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.351345 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.366175 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.377057 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.377132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.377148 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.377169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.377186 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.382859 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.397420 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.420061 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.446926 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.461757 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.479232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.479306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.479320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.479338 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.479375 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.480531 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.497066 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.513511 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.524934 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.536961 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.547668 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.562944 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.581132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.581167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.581178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.581196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.581208 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.584999 4758 generic.go:334] "Generic (PLEG): container finished" podID="af0a535d-83d3-4669-9245-fa99ddcb6a7d" containerID="519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f" exitCode=0 Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.585078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerDied","Data":"519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.594445 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.595333 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.595408 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.598999 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.622505 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.637557 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.640895 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.643348 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.654583 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.667851 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.681401 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.684445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.684478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.684489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.684510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.684524 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.701911 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.714495 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.726350 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.737692 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.746724 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.757478 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.769732 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.780748 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.786507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.786535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.786543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.786572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.786581 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.792642 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.806294 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.824324 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.837430 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.851003 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.862788 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.873848 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.888248 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.889033 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.889227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.889337 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.889439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.889523 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.898713 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.912132 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.926564 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.967267 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.987544 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.992254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.992294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.992303 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.992318 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:53 crc kubenswrapper[4758]: I1004 10:50:53.992326 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:53Z","lastTransitionTime":"2025-10-04T10:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.015880 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.095136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.095167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.095175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.095189 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.095197 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.197527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.197732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.197745 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.197761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.197773 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.300473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.300873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.300898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.300922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.300939 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.403693 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.403744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.403760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.403782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.403801 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.506450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.506499 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.506519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.506541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.506557 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.601017 4758 generic.go:334] "Generic (PLEG): container finished" podID="af0a535d-83d3-4669-9245-fa99ddcb6a7d" containerID="3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21" exitCode=0 Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.601074 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerDied","Data":"3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.601169 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.608720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.608758 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.608768 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.608782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.608792 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.621346 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.639259 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.659759 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.683904 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.711311 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.711617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.711782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.711903 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.712013 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.734270 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.753597 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.771038 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.791356 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.806173 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.813901 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.813928 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.813936 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.813948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.813957 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.819874 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.834913 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.849565 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.867717 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.882448 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.915746 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.915788 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.915799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.915810 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:54 crc kubenswrapper[4758]: I1004 10:50:54.915819 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:54Z","lastTransitionTime":"2025-10-04T10:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.018309 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.018347 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.018359 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.018375 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.018390 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.121248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.121301 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.121317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.121339 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.121355 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.224398 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.224445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.224455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.224469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.224480 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.325734 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.325819 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:55 crc kubenswrapper[4758]: E1004 10:50:55.325953 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.325978 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:55 crc kubenswrapper[4758]: E1004 10:50:55.326123 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:55 crc kubenswrapper[4758]: E1004 10:50:55.326235 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.328382 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.328421 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.328442 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.328470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.328493 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.432808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.433479 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.433617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.433759 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.433900 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.537012 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.537076 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.537095 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.537155 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.537179 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.613203 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" event={"ID":"af0a535d-83d3-4669-9245-fa99ddcb6a7d","Type":"ContainerStarted","Data":"178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.613318 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.635148 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.641863 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.641933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.641956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.641987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.642014 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.655872 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.672531 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.696530 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.730325 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.745044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.745140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.745162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.745187 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.745204 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.752441 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.772430 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.792084 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.808706 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.823665 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.846243 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.847645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.847706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.847725 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.847750 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.847769 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.876885 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.897555 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.920081 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.950757 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.950828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.950851 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.950884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:55 crc kubenswrapper[4758]: I1004 10:50:55.950915 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:55Z","lastTransitionTime":"2025-10-04T10:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.053645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.053710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.053728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.053756 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.053776 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.156844 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.156896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.156916 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.156945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.156967 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.261032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.261093 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.261134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.261158 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.261175 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.364583 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.364638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.364655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.364678 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.364694 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.467525 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.467599 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.467616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.467643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.467666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.570600 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.570657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.570676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.570702 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.570722 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.673697 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.673797 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.673815 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.673841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.673858 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.777654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.777713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.777733 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.777763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.777785 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.880648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.880857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.880883 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.880908 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.880924 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.983395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.983474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.983492 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.983513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:56 crc kubenswrapper[4758]: I1004 10:50:56.983554 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:56Z","lastTransitionTime":"2025-10-04T10:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.085846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.086050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.086071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.086090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.086123 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.188832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.188888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.188906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.188929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.188945 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.292895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.292958 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.292979 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.293008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.293030 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.325596 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.325662 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:57 crc kubenswrapper[4758]: E1004 10:50:57.325804 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:57 crc kubenswrapper[4758]: E1004 10:50:57.325876 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.326224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:57 crc kubenswrapper[4758]: E1004 10:50:57.326587 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.395568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.395662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.395683 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.395710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.395735 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.499015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.499072 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.499092 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.499204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.499229 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.602261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.602324 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.602341 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.602365 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.602384 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.705773 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.705825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.705841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.705865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.705882 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.809388 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.809444 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.809461 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.809484 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.809501 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.911976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.912012 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.912023 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.912039 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:57 crc kubenswrapper[4758]: I1004 10:50:57.912051 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:57Z","lastTransitionTime":"2025-10-04T10:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.014845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.014900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.014920 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.014943 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.014960 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.117417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.117455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.117464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.117480 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.117489 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.219674 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.219744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.219758 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.219774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.219811 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.323698 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.323729 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.323741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.323761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.323771 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.426464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.426504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.426515 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.426530 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.426541 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.528801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.528863 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.528883 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.528908 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.528927 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.585655 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.586062 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.601475 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" probeResult="failure" output="" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.621569 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" probeResult="failure" output="" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.631464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.631529 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.631550 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.631576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.631595 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.734003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.734040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.734052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.734070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.734130 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.835952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.835982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.835991 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.836004 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.836014 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.909548 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf"] Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.909974 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.911739 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.912224 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.932092 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.935547 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/84cf9e92-5435-48c3-8035-384e9455e6fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.935610 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/84cf9e92-5435-48c3-8035-384e9455e6fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.935632 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/84cf9e92-5435-48c3-8035-384e9455e6fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.935666 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wptfc\" (UniqueName: \"kubernetes.io/projected/84cf9e92-5435-48c3-8035-384e9455e6fd-kube-api-access-wptfc\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.939257 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.939282 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.939290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.939302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.939311 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:58Z","lastTransitionTime":"2025-10-04T10:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.949030 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.965615 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.977495 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:58 crc kubenswrapper[4758]: I1004 10:50:58.989603 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.006038 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.027071 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036418 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036467 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wptfc\" (UniqueName: \"kubernetes.io/projected/84cf9e92-5435-48c3-8035-384e9455e6fd-kube-api-access-wptfc\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036494 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036521 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/84cf9e92-5435-48c3-8035-384e9455e6fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036540 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/84cf9e92-5435-48c3-8035-384e9455e6fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036554 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/84cf9e92-5435-48c3-8035-384e9455e6fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.036579 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.036638 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.036677 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:15.036666432 +0000 UTC m=+52.329317321 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.036924 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:51:15.036916228 +0000 UTC m=+52.329567117 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.037114 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.037139 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:15.037132623 +0000 UTC m=+52.329783512 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.042527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.042565 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.042582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.042604 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.042663 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.042754 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/84cf9e92-5435-48c3-8035-384e9455e6fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.043075 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/84cf9e92-5435-48c3-8035-384e9455e6fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.043607 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/84cf9e92-5435-48c3-8035-384e9455e6fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.050022 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.068215 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.071360 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wptfc\" (UniqueName: \"kubernetes.io/projected/84cf9e92-5435-48c3-8035-384e9455e6fd-kube-api-access-wptfc\") pod \"ovnkube-control-plane-749d76644c-jx7sf\" (UID: \"84cf9e92-5435-48c3-8035-384e9455e6fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.082656 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.095675 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.110385 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.122991 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.135161 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.137925 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.138026 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138162 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138194 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138208 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138274 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:15.138250919 +0000 UTC m=+52.430901898 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138378 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138438 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138459 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.138554 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:15.138533396 +0000 UTC m=+52.431184315 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.145381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.145412 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.145423 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.145435 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.145445 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.150205 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.239675 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.247743 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.247793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.247809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.247835 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.247851 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: W1004 10:50:59.257745 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84cf9e92_5435_48c3_8035_384e9455e6fd.slice/crio-3a33814df774d0eda47a04bc33caf7253eae11b5eec3ee66e16a0d550d62293c WatchSource:0}: Error finding container 3a33814df774d0eda47a04bc33caf7253eae11b5eec3ee66e16a0d550d62293c: Status 404 returned error can't find the container with id 3a33814df774d0eda47a04bc33caf7253eae11b5eec3ee66e16a0d550d62293c Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.325397 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.325710 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.325812 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.325961 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.326201 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.326371 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.359297 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.359343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.359362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.359384 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.359401 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.462585 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.462622 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.462633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.462649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.462659 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.565673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.565710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.565721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.565764 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.565777 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.629218 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" event={"ID":"84cf9e92-5435-48c3-8035-384e9455e6fd","Type":"ContainerStarted","Data":"3a33814df774d0eda47a04bc33caf7253eae11b5eec3ee66e16a0d550d62293c"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.660278 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-srgqp"] Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.661239 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.661350 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.668737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.668778 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.668801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.668819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.668832 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.675266 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.691685 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.703230 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.714983 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.725662 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.738153 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.741954 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgwzr\" (UniqueName: \"kubernetes.io/projected/e5adcca6-45b0-46ad-ac95-57cce5121dc9-kube-api-access-lgwzr\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.742018 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.756198 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.768146 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.770782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.770802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.770813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.770828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.770840 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.782765 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.793822 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.808202 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.826773 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.843075 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.843171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgwzr\" (UniqueName: \"kubernetes.io/projected/e5adcca6-45b0-46ad-ac95-57cce5121dc9-kube-api-access-lgwzr\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.843256 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: E1004 10:50:59.843354 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:00.343326637 +0000 UTC m=+37.635977596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.844142 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.858340 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.873220 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgwzr\" (UniqueName: \"kubernetes.io/projected/e5adcca6-45b0-46ad-ac95-57cce5121dc9-kube-api-access-lgwzr\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.873842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.873915 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.873925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.873940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.874285 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.874138 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.890438 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:50:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.977283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.977318 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.977330 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.977346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.977357 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.995513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.995545 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.995557 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.995570 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:50:59 crc kubenswrapper[4758]: I1004 10:50:59.995580 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:50:59Z","lastTransitionTime":"2025-10-04T10:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.013598 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.018917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.018941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.018951 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.018967 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.018977 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.037543 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.042015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.042284 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.042428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.042568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.042715 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.063883 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.070388 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.070445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.070463 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.070487 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.070506 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.092877 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.100813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.100892 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.100917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.100946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.100976 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.311564 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.311893 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.313408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.313456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.313473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.313495 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.313513 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.350352 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.350566 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:00 crc kubenswrapper[4758]: E1004 10:51:00.350675 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:01.350650079 +0000 UTC m=+38.643301008 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.417407 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.417465 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.417481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.417504 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.417521 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.521672 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.521743 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.521784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.521816 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.521839 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.624535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.624607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.624631 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.624662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.624685 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.727828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.727879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.727896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.727921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.727938 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.836056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.836156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.836174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.836224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.836242 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.938839 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.938900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.938915 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.938937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:00 crc kubenswrapper[4758]: I1004 10:51:00.938953 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:00Z","lastTransitionTime":"2025-10-04T10:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.041471 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.041531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.041589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.041615 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.041638 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.143904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.143953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.143973 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.143997 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.144015 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.247222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.247266 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.247277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.247293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.247304 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.324803 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.324912 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.324990 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:01 crc kubenswrapper[4758]: E1004 10:51:01.325389 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:01 crc kubenswrapper[4758]: E1004 10:51:01.325016 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:01 crc kubenswrapper[4758]: E1004 10:51:01.325232 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.325041 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:01 crc kubenswrapper[4758]: E1004 10:51:01.325613 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.350061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.350173 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.350194 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.350218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.350237 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.360832 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:01 crc kubenswrapper[4758]: E1004 10:51:01.360998 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:01 crc kubenswrapper[4758]: E1004 10:51:01.361098 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:03.361068711 +0000 UTC m=+40.653719630 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.453993 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.454134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.454164 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.454195 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.454221 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.557005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.557063 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.557165 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.557196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.557215 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.640159 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/0.log" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.644220 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6" exitCode=1 Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.644331 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.646068 4758 scope.go:117] "RemoveContainer" containerID="f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.646685 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" event={"ID":"84cf9e92-5435-48c3-8035-384e9455e6fd","Type":"ContainerStarted","Data":"2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.646731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" event={"ID":"84cf9e92-5435-48c3-8035-384e9455e6fd","Type":"ContainerStarted","Data":"e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.661567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.661599 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.661611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.661625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.661636 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.666556 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.683426 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.701343 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.716641 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.733352 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.752006 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.765595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.765639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.765652 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.765670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.765681 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.772897 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.787330 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.805706 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.819430 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.841628 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.861367 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.867966 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.868008 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.868022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.868042 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.868059 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.877599 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.892815 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.910683 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.926597 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"ssIP event handler 8\\\\nI1004 10:50:59.182417 5970 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 10:50:59.182423 5970 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:50:59.182737 5970 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.182898 5970 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:50:59.183659 5970 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.184036 5970 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:50:59.184076 5970 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:50:59.184086 5970 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:50:59.184182 5970 factory.go:656] Stopping watch factory\\\\nI1004 10:50:59.184215 5970 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:50:59.184229 5970 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:50:59.184241 5970 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.938764 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.949936 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.961368 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.969901 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.969929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.969937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.969949 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.969957 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:01Z","lastTransitionTime":"2025-10-04T10:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.972024 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:01 crc kubenswrapper[4758]: I1004 10:51:01.986584 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.010646 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"ssIP event handler 8\\\\nI1004 10:50:59.182417 5970 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 10:50:59.182423 5970 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:50:59.182737 5970 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.182898 5970 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:50:59.183659 5970 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.184036 5970 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:50:59.184076 5970 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:50:59.184086 5970 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:50:59.184182 5970 factory.go:656] Stopping watch factory\\\\nI1004 10:50:59.184215 5970 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:50:59.184229 5970 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:50:59.184241 5970 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.021842 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.032836 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.050422 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.063866 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.072185 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.072218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.072230 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.072248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.072263 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.082167 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.095960 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.107000 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.122829 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.138529 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.152356 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.174166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.174218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.174234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.174258 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.174275 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.279275 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.279320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.279331 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.279351 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.279364 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.381967 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.382022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.382040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.382062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.382082 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.485073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.485150 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.485166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.485185 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.485199 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.587014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.587047 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.587056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.587067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.587076 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.652534 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/0.log" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.655669 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.656380 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.668453 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.684136 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.689257 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.689298 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.689309 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.689325 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.689338 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.695849 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.706324 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.720145 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.730740 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.762869 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.786055 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.791931 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.791962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.791970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.791982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.791991 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.817715 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.838232 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.858714 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.872020 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.884317 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.894811 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.894846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.894855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.894897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.894909 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.900456 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"ssIP event handler 8\\\\nI1004 10:50:59.182417 5970 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 10:50:59.182423 5970 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:50:59.182737 5970 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.182898 5970 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:50:59.183659 5970 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.184036 5970 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:50:59.184076 5970 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:50:59.184086 5970 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:50:59.184182 5970 factory.go:656] Stopping watch factory\\\\nI1004 10:50:59.184215 5970 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:50:59.184229 5970 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:50:59.184241 5970 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.913588 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.924281 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.997315 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.997370 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.997380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.997393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:02 crc kubenswrapper[4758]: I1004 10:51:02.997418 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:02Z","lastTransitionTime":"2025-10-04T10:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.099447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.099503 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.099522 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.099543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.099562 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.201770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.201846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.201868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.201897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.201920 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.304417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.304472 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.304489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.304513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.304531 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.325001 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.325255 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.325303 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.325462 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.325271 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.325539 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.325712 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.325869 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.342834 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.356804 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.374003 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.381529 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.381762 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.381856 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:07.381828125 +0000 UTC m=+44.674479074 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.388075 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.404917 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.406264 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.406300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.406314 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.406334 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.406348 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.419339 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.431866 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.445484 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.456545 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.467081 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.475026 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.495924 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.509431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.509469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.509478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.509493 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.509504 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.514047 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"ssIP event handler 8\\\\nI1004 10:50:59.182417 5970 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 10:50:59.182423 5970 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:50:59.182737 5970 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.182898 5970 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:50:59.183659 5970 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.184036 5970 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:50:59.184076 5970 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:50:59.184086 5970 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:50:59.184182 5970 factory.go:656] Stopping watch factory\\\\nI1004 10:50:59.184215 5970 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:50:59.184229 5970 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:50:59.184241 5970 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.524585 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.537439 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.552158 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.612202 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.612246 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.612259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.612275 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.612287 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.661780 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/1.log" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.662802 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/0.log" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.667248 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9" exitCode=1 Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.667345 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.667444 4758 scope.go:117] "RemoveContainer" containerID="f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.668332 4758 scope.go:117] "RemoveContainer" containerID="9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9" Oct 04 10:51:03 crc kubenswrapper[4758]: E1004 10:51:03.668546 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.687287 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.700919 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.714780 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.716096 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.716163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.716176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.716194 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.716209 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.731138 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.747039 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.761568 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.776748 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.791388 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.810009 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.820084 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.820169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.820188 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.820212 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.820231 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.824633 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.842054 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.856376 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.873284 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.895302 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93b6f27decefe55ebb4b5f2933765d10128531354f47d340cf35a33727c9ac6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:00Z\\\",\\\"message\\\":\\\"ssIP event handler 8\\\\nI1004 10:50:59.182417 5970 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 10:50:59.182423 5970 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:50:59.182737 5970 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.182898 5970 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:50:59.183659 5970 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:50:59.184036 5970 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:50:59.184076 5970 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:50:59.184086 5970 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:50:59.184182 5970 factory.go:656] Stopping watch factory\\\\nI1004 10:50:59.184215 5970 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:50:59.184229 5970 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:50:59.184241 5970 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:02Z\\\",\\\"message\\\":\\\"51:02.901407 6222 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:51:02.901720 6222 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.901874 6222 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.902021 6222 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902079 6222 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902282 6222 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902452 6222 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902516 6222 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902620 6222 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.911825 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.923410 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.923430 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.923438 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.923451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.923460 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:03Z","lastTransitionTime":"2025-10-04T10:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:03 crc kubenswrapper[4758]: I1004 10:51:03.924217 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:03Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.026168 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.026205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.026218 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.026235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.026247 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.128678 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.128727 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.128742 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.128763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.128778 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.231162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.231234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.231260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.231292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.231312 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.335363 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.335426 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.335457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.335481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.335497 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.438174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.438216 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.438226 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.438243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.438255 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.541346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.541418 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.541433 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.541450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.541465 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.643489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.643533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.643545 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.643561 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.643574 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.671438 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/1.log" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.674631 4758 scope.go:117] "RemoveContainer" containerID="9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9" Oct 04 10:51:04 crc kubenswrapper[4758]: E1004 10:51:04.674788 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.692807 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.709295 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.727888 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.745148 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.747399 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.747454 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.747478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.747506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.747528 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.766383 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.786960 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.803923 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.824773 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.845810 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.850433 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.850472 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.850511 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.850531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.850542 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.865429 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.881021 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.903017 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.931964 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:02Z\\\",\\\"message\\\":\\\"51:02.901407 6222 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:51:02.901720 6222 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.901874 6222 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.902021 6222 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902079 6222 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902282 6222 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902452 6222 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902516 6222 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902620 6222 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.948360 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.952914 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.952970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.953006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.953022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.953033 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:04Z","lastTransitionTime":"2025-10-04T10:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.964727 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:04 crc kubenswrapper[4758]: I1004 10:51:04.985748 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.055959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.056036 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.056054 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.056080 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.056127 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.159083 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.159238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.159261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.159285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.159303 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.262712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.262799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.262817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.262875 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.262894 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.324945 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.324974 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:05 crc kubenswrapper[4758]: E1004 10:51:05.325158 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.325705 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:05 crc kubenswrapper[4758]: E1004 10:51:05.325820 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.325868 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:05 crc kubenswrapper[4758]: E1004 10:51:05.325959 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:05 crc kubenswrapper[4758]: E1004 10:51:05.326036 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.366784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.366874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.366900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.366934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.366957 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.469941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.470010 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.470025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.470045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.470056 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.573389 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.573448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.573464 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.573488 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.573504 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.676715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.676779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.676792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.676809 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.676819 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.778990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.779045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.779062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.779085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.779133 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.881628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.881676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.881693 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.881714 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.881730 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.984970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.985048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.985070 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.985097 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:05 crc kubenswrapper[4758]: I1004 10:51:05.985149 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:05Z","lastTransitionTime":"2025-10-04T10:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.088266 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.088367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.088387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.088411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.088426 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.191706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.191806 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.191823 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.191847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.191865 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.294559 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.294626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.294643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.294667 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.294684 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.396822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.396876 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.396897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.396925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.396945 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.499381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.499438 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.499451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.499468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.499479 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.602616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.602666 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.602682 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.602704 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.602721 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.707312 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.707706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.707723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.707748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.707765 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.810728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.810774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.810792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.810813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.810830 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.913474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.913522 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.913543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.913565 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:06 crc kubenswrapper[4758]: I1004 10:51:06.913582 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:06Z","lastTransitionTime":"2025-10-04T10:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.016599 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.016675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.016699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.016729 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.016753 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.119877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.119930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.119952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.119985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.120007 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.223262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.223331 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.223351 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.223379 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.223433 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.325011 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.325078 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.325015 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:07 crc kubenswrapper[4758]: E1004 10:51:07.325267 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.325304 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:07 crc kubenswrapper[4758]: E1004 10:51:07.325427 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:07 crc kubenswrapper[4758]: E1004 10:51:07.325629 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:07 crc kubenswrapper[4758]: E1004 10:51:07.325806 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.327325 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.327404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.327426 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.327457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.327481 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.431355 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:07 crc kubenswrapper[4758]: E1004 10:51:07.431571 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:07 crc kubenswrapper[4758]: E1004 10:51:07.431679 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:15.43165449 +0000 UTC m=+52.724305419 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.432954 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.433167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.433306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.433497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.433629 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.536529 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.536839 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.536983 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.537156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.537294 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.641783 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.641835 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.641851 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.641874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.641896 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.745355 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.745414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.745434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.745457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.745474 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.848228 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.848280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.848297 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.848319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.848336 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.951175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.951251 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.951272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.951296 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:07 crc kubenswrapper[4758]: I1004 10:51:07.951315 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:07Z","lastTransitionTime":"2025-10-04T10:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.054061 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.054153 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.054172 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.054239 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.054258 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.157613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.157986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.158175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.158312 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.158429 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.261201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.261271 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.261294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.261323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.261341 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.364374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.364434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.364451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.364587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.364653 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.467845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.467900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.467918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.467944 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.467960 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.570869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.570925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.570941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.570963 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.570980 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.672893 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.672938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.672945 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.672957 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.672965 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.776145 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.776196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.776213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.776238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.776255 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.878942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.879002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.879018 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.879041 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.879058 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.982365 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.982444 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.982476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.982510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:08 crc kubenswrapper[4758]: I1004 10:51:08.982530 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:08Z","lastTransitionTime":"2025-10-04T10:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.085711 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.085770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.085788 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.085811 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.085830 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.188346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.188395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.188410 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.188431 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.188449 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.291183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.291225 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.291236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.291253 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.291266 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.326519 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.326554 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.326581 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.326675 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:09 crc kubenswrapper[4758]: E1004 10:51:09.326671 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:09 crc kubenswrapper[4758]: E1004 10:51:09.326806 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:09 crc kubenswrapper[4758]: E1004 10:51:09.326875 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:09 crc kubenswrapper[4758]: E1004 10:51:09.326963 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.393774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.393826 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.393842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.393865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.393882 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.496631 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.496670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.496684 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.496701 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.496713 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.599476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.599535 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.599553 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.599577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.599594 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.702140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.702178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.702186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.702201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.702210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.804646 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.804691 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.804708 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.804722 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.804732 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.907049 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.907291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.907304 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.907319 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:09 crc kubenswrapper[4758]: I1004 10:51:09.907328 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:09Z","lastTransitionTime":"2025-10-04T10:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.009614 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.009706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.009723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.009783 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.009797 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.112418 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.112681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.112782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.112885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.112913 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.215932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.215973 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.215985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.216012 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.216025 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.319059 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.319086 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.319094 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.319127 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.319138 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.421660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.421702 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.421715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.421732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.421749 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.525062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.525132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.525149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.525260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.525291 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.627870 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.627904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.627929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.627942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.627952 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.695830 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.695877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.695888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.695902 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.695913 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: E1004 10:51:10.714507 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:10Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.720441 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.720490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.720512 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.720539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.720560 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: E1004 10:51:10.748226 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:10Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.752653 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.752694 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.752705 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.752720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.752734 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: E1004 10:51:10.771136 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:10Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.777422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.777472 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.777490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.777515 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.777533 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: E1004 10:51:10.795917 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:10Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.801804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.801861 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.801879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.801902 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.801921 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: E1004 10:51:10.823213 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:10Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:10 crc kubenswrapper[4758]: E1004 10:51:10.823342 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.832630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.832669 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.832679 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.832696 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.832706 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.935562 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.935855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.935961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.936066 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:10 crc kubenswrapper[4758]: I1004 10:51:10.936273 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:10Z","lastTransitionTime":"2025-10-04T10:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.039041 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.039126 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.039144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.039166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.039183 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.141157 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.141198 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.141206 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.141220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.141229 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.243886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.243913 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.243922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.243936 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.243947 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.325152 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.325182 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.325245 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:11 crc kubenswrapper[4758]: E1004 10:51:11.325284 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:11 crc kubenswrapper[4758]: E1004 10:51:11.325422 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.325440 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:11 crc kubenswrapper[4758]: E1004 10:51:11.325717 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:11 crc kubenswrapper[4758]: E1004 10:51:11.325622 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.345572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.345764 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.345863 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.345959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.346059 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.448247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.448283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.448291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.448302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.448312 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.551214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.551256 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.551272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.551292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.551307 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.653978 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.654435 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.654586 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.654724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.654855 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.757775 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.758193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.758349 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.758503 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.758630 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.861626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.861959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.862161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.862335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.862478 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.966346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.967779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.967942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.968081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:11 crc kubenswrapper[4758]: I1004 10:51:11.968280 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:11Z","lastTransitionTime":"2025-10-04T10:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.071484 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.071550 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.071571 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.071602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.071622 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.174995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.175200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.175230 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.175260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.175282 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.277567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.277629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.277648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.277673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.277690 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.379914 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.379971 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.379988 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.380013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.380029 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.481873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.481918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.481930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.481946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.481960 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.584799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.584914 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.584932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.584955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.584973 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.687956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.688014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.688030 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.688053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.688069 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.791207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.791273 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.791295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.791323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.791345 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.894427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.894479 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.894495 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.894518 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.894535 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.997674 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.997728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.997744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.997767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:12 crc kubenswrapper[4758]: I1004 10:51:12.997784 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:12Z","lastTransitionTime":"2025-10-04T10:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.100675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.100744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.100765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.100795 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.100816 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.203553 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.203609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.203626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.203654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.203672 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.306016 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.306392 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.306455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.306491 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.306513 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.325471 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:13 crc kubenswrapper[4758]: E1004 10:51:13.325928 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.326918 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:13 crc kubenswrapper[4758]: E1004 10:51:13.327087 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.327232 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:13 crc kubenswrapper[4758]: E1004 10:51:13.327354 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.327452 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:13 crc kubenswrapper[4758]: E1004 10:51:13.327574 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.352404 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.370853 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.394359 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.412552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.412632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.412650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.412675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.412697 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.426210 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:02Z\\\",\\\"message\\\":\\\"51:02.901407 6222 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:51:02.901720 6222 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.901874 6222 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.902021 6222 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902079 6222 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902282 6222 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902452 6222 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902516 6222 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902620 6222 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.447478 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.467548 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.485315 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.499571 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.515045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.515087 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.515135 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.515199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.515218 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.518742 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.535055 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.553016 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.569828 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.590832 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.618154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.618216 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.618235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.618259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.618287 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.619067 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.642679 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.662208 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:13Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.721073 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.721167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.721186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.721207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.721223 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.824066 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.824152 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.824171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.824191 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.824206 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.927084 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.927190 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.927245 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.927268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:13 crc kubenswrapper[4758]: I1004 10:51:13.927283 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:13Z","lastTransitionTime":"2025-10-04T10:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.029958 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.030019 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.030034 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.030055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.030074 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.132698 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.132754 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.132770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.132791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.132808 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.235704 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.235739 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.235748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.235763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.235774 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.338581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.338617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.338626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.338639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.338649 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.441265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.441346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.441371 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.441405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.441428 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.547450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.547506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.547524 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.547548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.547567 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.650977 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.651041 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.651064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.651092 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.651157 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.754541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.754609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.754633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.754662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.754685 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.857241 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.857277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.857288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.857305 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.857316 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.959735 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.959808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.959832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.959858 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:14 crc kubenswrapper[4758]: I1004 10:51:14.959875 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:14Z","lastTransitionTime":"2025-10-04T10:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.061990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.062048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.062065 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.062092 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.062138 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.111225 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.111371 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.111485 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:51:47.111449341 +0000 UTC m=+84.404100260 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.111557 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.111576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.111644 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:47.111625525 +0000 UTC m=+84.404276424 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.111780 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.111912 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:47.111882331 +0000 UTC m=+84.404533270 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.165255 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.165353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.165370 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.165395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.165415 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.212144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.212191 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212405 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212432 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212434 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212486 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212508 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212446 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212585 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:47.212561977 +0000 UTC m=+84.505212906 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.212618 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:47.212598158 +0000 UTC m=+84.505249117 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.267872 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.267919 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.267933 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.267951 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.267965 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.325391 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.325448 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.325391 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.325635 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.325968 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.326069 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.326280 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.326423 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.370372 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.370414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.370423 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.370436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.370446 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.474179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.474246 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.474263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.474455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.474474 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.516047 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.516246 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: E1004 10:51:15.516344 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:51:31.516326525 +0000 UTC m=+68.808977414 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.577989 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.578028 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.578040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.578059 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.578071 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.680336 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.680385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.680400 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.680419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.680434 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.783556 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.783614 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.783629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.783654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.783671 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.886726 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.886767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.886779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.886832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.886846 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.989649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.989692 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.989704 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.989720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:15 crc kubenswrapper[4758]: I1004 10:51:15.989732 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:15Z","lastTransitionTime":"2025-10-04T10:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.094264 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.094314 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.094330 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.094352 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.094368 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.197554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.197647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.197674 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.197710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.197733 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.301568 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.301625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.301643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.301667 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.301687 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.326518 4758 scope.go:117] "RemoveContainer" containerID="9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.406066 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.406169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.406187 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.406215 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.406233 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.508791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.508852 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.508869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.508894 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.508911 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.611633 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.611696 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.611713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.611737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.611754 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.715540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.715591 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.715607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.715630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.715647 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.725811 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/1.log" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.729712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.730369 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.751766 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.768573 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.780949 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.805064 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.816948 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.818740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.818795 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.818804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.818819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.818829 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.827439 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.840032 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.852948 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.868994 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.883177 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.905926 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.921125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.921149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.921156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.921170 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.921178 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:16Z","lastTransitionTime":"2025-10-04T10:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.926436 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:02Z\\\",\\\"message\\\":\\\"51:02.901407 6222 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:51:02.901720 6222 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.901874 6222 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.902021 6222 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902079 6222 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902282 6222 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902452 6222 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902516 6222 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902620 6222 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.937911 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.948193 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.959705 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:16 crc kubenswrapper[4758]: I1004 10:51:16.971325 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:16Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.023765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.023827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.023842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.023865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.023887 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.126796 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.126872 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.126891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.126926 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.126943 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.229887 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.229929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.229940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.229956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.229967 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.325536 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.325610 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.325637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.325667 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:17 crc kubenswrapper[4758]: E1004 10:51:17.325928 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:17 crc kubenswrapper[4758]: E1004 10:51:17.326166 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:17 crc kubenswrapper[4758]: E1004 10:51:17.326346 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:17 crc kubenswrapper[4758]: E1004 10:51:17.326546 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.332154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.332211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.332232 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.332260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.332278 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.435613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.435668 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.435685 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.435712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.435730 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.539370 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.539428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.539445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.539469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.539488 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.641976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.642037 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.642055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.642080 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.642140 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.737174 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/2.log" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.738136 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/1.log" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.742404 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7" exitCode=1 Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.742457 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.742503 4758 scope.go:117] "RemoveContainer" containerID="9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.745877 4758 scope.go:117] "RemoveContainer" containerID="49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7" Oct 04 10:51:17 crc kubenswrapper[4758]: E1004 10:51:17.746335 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.749707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.749776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.749799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.749828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.749850 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.777293 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.797763 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.810745 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.856038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.856136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.856159 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.856184 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.856200 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.868066 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.885991 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.902205 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.923157 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.943529 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.959211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.959268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.959285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.959308 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.959326 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:17Z","lastTransitionTime":"2025-10-04T10:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.960282 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:17 crc kubenswrapper[4758]: I1004 10:51:17.981996 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:17Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.007506 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.026475 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.062938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.063000 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.063017 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.063042 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.063059 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.079033 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.124321 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.147545 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:02Z\\\",\\\"message\\\":\\\"51:02.901407 6222 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:51:02.901720 6222 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.901874 6222 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.902021 6222 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902079 6222 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902282 6222 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902452 6222 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902516 6222 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902620 6222 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.159883 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.165788 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.166143 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.166159 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.166176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.166185 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.268677 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.268714 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.268723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.268737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.268747 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.371195 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.371250 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.371267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.371289 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.371305 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.474865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.474914 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.474925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.474944 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.474957 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.577952 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.578226 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.578291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.578352 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.578447 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.668411 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.681453 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.682440 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.682531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.682550 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.682611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.682636 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.685154 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.700711 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.716022 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.736857 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.748153 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/2.log" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.752856 4758 scope.go:117] "RemoveContainer" containerID="49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7" Oct 04 10:51:18 crc kubenswrapper[4758]: E1004 10:51:18.753134 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.760589 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.775214 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.784574 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.784647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.784658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.784671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.784680 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.795305 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.817828 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.833167 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.846004 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.860538 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.873514 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.886812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.886854 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.886869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.886888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.886902 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.891495 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.911853 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9fccaa67d253a2b0d283e28668e8c15f6191a2efd22421ee15b003f93d0976c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:02Z\\\",\\\"message\\\":\\\"51:02.901407 6222 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:51:02.901720 6222 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.901874 6222 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:51:02.902021 6222 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902079 6222 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902282 6222 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902452 6222 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902516 6222 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:51:02.902620 6222 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.926577 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.938989 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.954279 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.966262 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.981892 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.989571 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.989612 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.989625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.989645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:18 crc kubenswrapper[4758]: I1004 10:51:18.989658 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:18Z","lastTransitionTime":"2025-10-04T10:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.006126 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.019288 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.034684 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.048418 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.058820 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.068115 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.080187 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.089287 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.092002 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.092051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.092062 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.092077 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.092088 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.099639 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.109402 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.123793 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.136073 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.146873 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.157453 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.194429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.194486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.194502 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.194594 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.194615 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.297029 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.297077 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.297091 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.297141 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.297156 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.324693 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.324745 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.324746 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:19 crc kubenswrapper[4758]: E1004 10:51:19.324791 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:19 crc kubenswrapper[4758]: E1004 10:51:19.324854 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.324879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:19 crc kubenswrapper[4758]: E1004 10:51:19.324973 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:19 crc kubenswrapper[4758]: E1004 10:51:19.325013 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.399953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.399992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.400001 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.400016 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.400027 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.502834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.502878 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.502890 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.502904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.502915 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.605818 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.605893 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.605908 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.605936 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.605953 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.708977 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.709032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.709044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.709063 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.709079 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.811771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.811856 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.811877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.811909 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.811928 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.914644 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.914691 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.914703 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.914720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:19 crc kubenswrapper[4758]: I1004 10:51:19.914730 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:19Z","lastTransitionTime":"2025-10-04T10:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.016721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.016759 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.016766 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.016779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.016788 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.119032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.119089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.119122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.119141 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.119152 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.221352 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.221394 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.221403 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.221420 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.221430 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.323210 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.323268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.323277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.323291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.323300 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.425637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.425737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.425749 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.425761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.425769 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.528248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.528280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.528288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.528301 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.528309 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.631060 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.631125 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.631145 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.631161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.631172 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.733586 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.733620 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.733629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.733641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.733650 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.836194 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.836238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.836251 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.836269 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.836282 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.939227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.939267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.939277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.939291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.939301 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.949907 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.949979 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.949995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.950019 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.950036 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: E1004 10:51:20.969009 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.972591 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.972640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.972655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.972676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.972693 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:20 crc kubenswrapper[4758]: E1004 10:51:20.985021 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.988650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.988695 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.988710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.988732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:20 crc kubenswrapper[4758]: I1004 10:51:20.988747 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:20Z","lastTransitionTime":"2025-10-04T10:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.005567 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:21Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.009545 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.009626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.009639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.009656 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.009697 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.026911 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:21Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.030301 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.030338 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.030349 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.030364 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.030378 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.048083 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:21Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.048346 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.050016 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.050071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.050090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.050139 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.050160 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.153321 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.153607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.153721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.153813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.153908 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.256700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.256737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.256744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.256760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.256768 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.326255 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.326332 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.326425 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.326492 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.326504 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.326644 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.326714 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:21 crc kubenswrapper[4758]: E1004 10:51:21.326914 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.358796 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.358855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.358871 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.358896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.358912 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.460739 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.460801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.460819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.460843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.460860 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.562917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.562959 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.562967 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.562980 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.562988 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.666406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.666456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.666473 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.666496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.666513 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.769706 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.769766 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.769782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.769806 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.769823 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.873181 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.873241 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.873262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.873291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.873313 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.976966 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.977026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.977045 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.977072 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:21 crc kubenswrapper[4758]: I1004 10:51:21.977091 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:21Z","lastTransitionTime":"2025-10-04T10:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.079247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.079274 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.079281 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.079294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.079302 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.181480 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.181527 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.181538 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.181554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.181563 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.284129 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.284180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.284192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.284209 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.284221 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.387439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.388044 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.388163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.388264 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.388395 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.491591 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.491628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.491637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.491650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.491659 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.594355 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.594418 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.594434 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.594456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.594474 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.696618 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.696650 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.696658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.696673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.696681 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.798267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.798506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.798581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.798647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.798704 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.900597 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.900676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.900693 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.900717 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:22 crc kubenswrapper[4758]: I1004 10:51:22.900738 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:22Z","lastTransitionTime":"2025-10-04T10:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.002678 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.002889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.002982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.003049 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.003118 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.105214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.105260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.105272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.105292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.105305 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.207906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.208154 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.208235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.208316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.208388 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.311348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.311411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.311428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.311451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.311470 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.327300 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:23 crc kubenswrapper[4758]: E1004 10:51:23.327551 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.327933 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.327956 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:23 crc kubenswrapper[4758]: E1004 10:51:23.328147 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.328225 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:23 crc kubenswrapper[4758]: E1004 10:51:23.328404 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:23 crc kubenswrapper[4758]: E1004 10:51:23.328576 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.354079 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.373653 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.393385 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.407006 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.417475 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.417563 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.417589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.418080 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.418482 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.426607 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.455184 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.472288 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.484471 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.497133 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.513522 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.521704 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.521748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.521761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.521778 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.521791 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.538650 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.551168 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.564070 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.575085 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.594245 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.609997 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.622585 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.624354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.624404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.624416 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.624433 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.624679 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.727566 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.727608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.727621 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.727638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.727650 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.831180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.831281 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.831306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.831336 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.831419 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.936064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.936171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.936196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.936224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:23 crc kubenswrapper[4758]: I1004 10:51:23.936244 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:23Z","lastTransitionTime":"2025-10-04T10:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.038513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.038570 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.038587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.038613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.038630 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.141259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.141613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.141756 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.141899 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.142084 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.245873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.245922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.245938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.245962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.245980 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.349381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.349677 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.349887 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.350082 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.350408 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.453405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.454815 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.455071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.455367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.455557 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.558661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.558708 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.558720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.558737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.558748 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.661763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.661917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.661939 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.661967 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.661987 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.764506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.764592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.764608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.764623 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.764634 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.866822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.866850 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.866859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.866873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.866885 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.969912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.969978 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.970021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.970053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:24 crc kubenswrapper[4758]: I1004 10:51:24.970075 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:24Z","lastTransitionTime":"2025-10-04T10:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.072860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.072918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.072937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.072961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.072977 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.175931 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.176005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.176025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.176053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.176078 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.279005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.279058 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.279074 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.279120 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.279137 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.325898 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.326095 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.326137 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:25 crc kubenswrapper[4758]: E1004 10:51:25.326480 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.326515 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:25 crc kubenswrapper[4758]: E1004 10:51:25.326721 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:25 crc kubenswrapper[4758]: E1004 10:51:25.326637 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:25 crc kubenswrapper[4758]: E1004 10:51:25.327158 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.380737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.380827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.380843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.380866 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.380883 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.483837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.483907 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.483915 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.483930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.483941 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.586739 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.586793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.586803 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.586820 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.586832 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.689490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.689543 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.689560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.689586 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.689606 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.793344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.793411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.793429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.793453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.793472 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.896956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.897009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.897026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.897049 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.897066 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.999770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.999817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:25 crc kubenswrapper[4758]: I1004 10:51:25.999828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:25.999845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:25.999857 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:25Z","lastTransitionTime":"2025-10-04T10:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.101964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.102031 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.102053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.102081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.102147 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.204456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.204510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.204526 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.204548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.204564 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.309056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.309268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.309337 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.309373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.309414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.411911 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.412247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.412317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.412380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.412440 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.514860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.515847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.516613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.516851 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.517029 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.619617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.619712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.619767 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.619790 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.619809 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.743719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.743773 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.743791 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.743814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.743832 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.846437 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.846522 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.846540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.846563 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.846579 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.949162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.949199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.949211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.949227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:26 crc kubenswrapper[4758]: I1004 10:51:26.949239 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:26Z","lastTransitionTime":"2025-10-04T10:51:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.051260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.051338 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.051356 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.051381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.051399 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.154067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.154158 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.154178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.154201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.154218 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.257360 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.257408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.257423 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.257442 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.257455 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.325406 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:27 crc kubenswrapper[4758]: E1004 10:51:27.325703 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.325456 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:27 crc kubenswrapper[4758]: E1004 10:51:27.325971 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.325423 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:27 crc kubenswrapper[4758]: E1004 10:51:27.326172 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.325523 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:27 crc kubenswrapper[4758]: E1004 10:51:27.326251 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.360201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.360277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.360294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.360314 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.360330 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.463316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.463663 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.463683 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.463705 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.463721 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.565805 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.566450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.566542 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.566632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.566726 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.669961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.669998 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.670015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.670035 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.670052 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.772343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.772393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.772408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.772428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.772445 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.876041 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.876219 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.876324 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.876422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.876516 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.978366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.978514 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.978585 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.978649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:27 crc kubenswrapper[4758]: I1004 10:51:27.978819 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:27Z","lastTransitionTime":"2025-10-04T10:51:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.080772 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.081180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.081350 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.081495 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.081569 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.184827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.184874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.184889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.184906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.184918 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.287625 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.287674 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.287683 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.287699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.287710 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.390092 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.390171 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.390183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.390204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.390218 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.493144 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.493193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.493205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.493231 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.493244 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.596302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.596351 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.596363 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.596402 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.596414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.698448 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.698619 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.698643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.698671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.698773 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.800635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.800679 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.800692 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.800714 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.800728 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.902763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.902810 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.902822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.902840 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:28 crc kubenswrapper[4758]: I1004 10:51:28.902852 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:28Z","lastTransitionTime":"2025-10-04T10:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.005119 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.005261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.005327 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.005405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.005468 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.107462 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.107647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.107731 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.107795 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.107854 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.210724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.211011 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.211122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.211213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.211283 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.313597 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.313632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.313640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.313655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.313664 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.325051 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.325058 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.325166 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.325212 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:29 crc kubenswrapper[4758]: E1004 10:51:29.325290 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:29 crc kubenswrapper[4758]: E1004 10:51:29.325413 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:29 crc kubenswrapper[4758]: E1004 10:51:29.325528 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:29 crc kubenswrapper[4758]: E1004 10:51:29.325589 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.416290 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.416322 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.416333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.416348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.416358 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.519339 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.519393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.519405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.519426 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.519438 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.622511 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.622578 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.622591 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.622652 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.622666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.724681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.724725 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.724770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.724786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.724796 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.826696 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.826765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.826776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.826789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.826798 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.930361 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.930450 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.930463 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.930481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:29 crc kubenswrapper[4758]: I1004 10:51:29.930492 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:29Z","lastTransitionTime":"2025-10-04T10:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.033839 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.033891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.033904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.033922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.033937 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.136476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.136530 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.136540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.136554 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.136564 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.238872 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.238925 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.238937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.238955 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.238969 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.326889 4758 scope.go:117] "RemoveContainer" containerID="49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7" Oct 04 10:51:30 crc kubenswrapper[4758]: E1004 10:51:30.327215 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.336791 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.341768 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.341817 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.341833 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.341859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.341874 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.444793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.444859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.444880 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.444898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.444915 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.548138 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.548258 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.548333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.548401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.548476 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.651911 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.652332 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.652414 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.652481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.652546 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.755605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.755680 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.755710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.755748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.755772 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.858466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.858521 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.858534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.858548 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.858558 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.960976 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.961167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.961243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.961313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:30 crc kubenswrapper[4758]: I1004 10:51:30.961375 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:30Z","lastTransitionTime":"2025-10-04T10:51:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.064711 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.064752 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.064760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.064775 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.064783 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.167470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.167537 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.167551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.167572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.167588 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.269819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.269869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.269881 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.269897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.269908 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.325498 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.325525 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.325624 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.325637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.325511 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.325748 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.325813 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.325866 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.372629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.372676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.372688 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.372705 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.372718 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.384740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.384789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.384804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.384819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.385210 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.396272 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:31Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.400288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.400327 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.400338 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.400354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.400369 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.412969 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:31Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.416427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.416463 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.416474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.416491 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.416502 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.427029 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:31Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.430492 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.430517 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.430525 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.430538 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.430548 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.440054 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:31Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.442951 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.443055 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.443142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.443225 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.443289 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.454923 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:31Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.455130 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.474537 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.474641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.474707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.474778 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.474836 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.576469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.576493 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.576501 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.576512 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.576520 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.595212 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.595365 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:31 crc kubenswrapper[4758]: E1004 10:51:31.595477 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:52:03.595462737 +0000 UTC m=+100.888113626 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.678552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.678612 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.678622 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.678635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.678643 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.780799 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.780850 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.780860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.780873 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.780882 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.882974 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.883050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.883071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.883097 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.883151 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.985808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.985870 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.985886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.985909 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:31 crc kubenswrapper[4758]: I1004 10:51:31.985929 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:31Z","lastTransitionTime":"2025-10-04T10:51:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.088128 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.088177 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.088190 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.088205 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.088214 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.190267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.190307 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.190316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.190328 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.190337 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.292474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.292502 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.292526 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.292540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.292548 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.394295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.394338 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.394346 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.394359 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.394368 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.496572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.496614 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.496628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.496647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.496659 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.598655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.598741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.598749 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.598763 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.598774 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.700776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.700810 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.700818 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.700834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.700843 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.802519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.802552 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.802560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.802572 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.802581 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.905801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.905859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.905877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.905903 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:32 crc kubenswrapper[4758]: I1004 10:51:32.905922 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:32Z","lastTransitionTime":"2025-10-04T10:51:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.008042 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.008071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.008079 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.008094 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.008126 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.110753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.110779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.110787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.110798 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.110807 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.212613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.212635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.212643 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.212654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.212662 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.315212 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.315261 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.315279 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.315300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.315319 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.325283 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.325316 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.325317 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.325285 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:33 crc kubenswrapper[4758]: E1004 10:51:33.325372 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:33 crc kubenswrapper[4758]: E1004 10:51:33.325439 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:33 crc kubenswrapper[4758]: E1004 10:51:33.325496 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:33 crc kubenswrapper[4758]: E1004 10:51:33.325538 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.335767 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.350692 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.360153 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.370286 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.384418 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.398659 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.411618 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.422211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.422234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.422242 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.422254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.422263 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.422721 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.435250 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.459607 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.472328 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.481750 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.492294 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.503705 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.515917 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.524306 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.524410 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.524468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.524549 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.524618 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.527125 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.538822 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.549144 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:33Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.627173 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.627223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.627239 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.627263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.627281 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.729200 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.729263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.729283 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.729311 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.729329 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.832699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.832771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.832789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.832864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.832882 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.935362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.935394 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.935404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.935417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:33 crc kubenswrapper[4758]: I1004 10:51:33.935425 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:33Z","lastTransitionTime":"2025-10-04T10:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.037721 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.037769 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.037777 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.037789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.037797 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.139779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.139820 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.139828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.139842 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.139851 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.242119 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.242148 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.242156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.242168 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.242176 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.343631 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.343670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.343679 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.343697 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.343707 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.445981 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.446030 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.446046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.446068 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.446085 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.548616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.548640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.548648 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.548659 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.548667 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.650640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.650703 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.650712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.650723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.650731 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.752987 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.753014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.753022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.753034 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.753043 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.803688 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/0.log" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.803721 4758 generic.go:334] "Generic (PLEG): container finished" podID="3875fe1d-de8c-4266-bf2b-e07c633b85dc" containerID="13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1" exitCode=1 Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.803742 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerDied","Data":"13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.804018 4758 scope.go:117] "RemoveContainer" containerID="13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.817707 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.830266 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.840208 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.852053 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.856236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.856272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.856281 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.856307 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.856316 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.863276 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.872237 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.883508 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.894541 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.907388 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.918187 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.930189 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.944606 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.954149 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.958391 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.958424 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.958433 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.958447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.958457 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:34Z","lastTransitionTime":"2025-10-04T10:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.965526 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.975658 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.987576 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:34 crc kubenswrapper[4758]: I1004 10:51:34.998565 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:34Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.009655 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.060545 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.060570 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.060578 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.060592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.060604 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.162417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.162457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.162465 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.162479 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.162489 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.265003 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.265046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.265056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.265072 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.265082 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.325659 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.325666 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:35 crc kubenswrapper[4758]: E1004 10:51:35.325782 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.325815 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.325879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:35 crc kubenswrapper[4758]: E1004 10:51:35.325967 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:35 crc kubenswrapper[4758]: E1004 10:51:35.326038 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:35 crc kubenswrapper[4758]: E1004 10:51:35.326122 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.367427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.367466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.367475 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.367489 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.367498 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.469169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.469220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.469235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.469255 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.469267 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.571167 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.571212 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.571223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.571240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.571253 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.673190 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.673227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.673236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.673259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.673269 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.775332 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.775371 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.775381 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.775395 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.775404 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.808742 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/0.log" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.808793 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerStarted","Data":"b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.824000 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.833802 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.851126 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.873632 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.877166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.877236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.877254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.877280 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.877298 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.888292 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.907727 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.922525 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.931865 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.939835 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.949051 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.958218 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.971318 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.979364 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.979400 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.979408 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.979421 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.979430 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:35Z","lastTransitionTime":"2025-10-04T10:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:35 crc kubenswrapper[4758]: I1004 10:51:35.983299 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:35Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.001808 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:36Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.016194 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:36Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.024703 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:36Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.034648 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:36Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.044509 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:36Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.080854 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.080882 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.080921 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.080934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.080944 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.183244 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.183317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.183335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.183358 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.183375 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.285297 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.285362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.285380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.285405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.285422 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.387725 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.387789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.387808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.387833 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.387868 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.490428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.490468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.490476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.490490 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.490500 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.592607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.592641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.592649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.592661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.592669 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.694449 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.694477 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.694486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.694498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.694506 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.797453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.797488 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.797499 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.797513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.797526 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.899189 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.899237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.899248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.899265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:36 crc kubenswrapper[4758]: I1004 10:51:36.899276 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:36Z","lastTransitionTime":"2025-10-04T10:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.001699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.001770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.001788 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.001812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.001830 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.104942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.104992 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.105009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.105032 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.105051 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.208036 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.208085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.208130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.208156 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.208185 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.310608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.310645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.310654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.310669 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.310680 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.325324 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.325350 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.325365 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:37 crc kubenswrapper[4758]: E1004 10:51:37.325423 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.325585 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:37 crc kubenswrapper[4758]: E1004 10:51:37.325599 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:37 crc kubenswrapper[4758]: E1004 10:51:37.325642 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:37 crc kubenswrapper[4758]: E1004 10:51:37.325713 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.413454 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.413497 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.413511 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.413533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.413549 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.515624 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.515649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.515660 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.515673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.515686 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.618614 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.618649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.618663 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.618678 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.618689 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.720761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.720859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.720917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.720938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.720949 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.824516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.824587 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.824613 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.824641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.824666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.927635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.927675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.927683 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.927698 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:37 crc kubenswrapper[4758]: I1004 10:51:37.927707 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:37Z","lastTransitionTime":"2025-10-04T10:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.031386 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.031451 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.031472 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.031500 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.031521 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.134317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.134385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.134406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.134437 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.134504 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.237474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.237524 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.237534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.237560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.237569 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.340478 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.340551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.340573 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.340601 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.340623 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.443729 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.443775 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.443786 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.443807 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.443819 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.546585 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.546647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.546659 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.546696 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.546711 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.649627 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.649671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.649681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.649696 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.649708 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.752316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.752347 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.752354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.752366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.752375 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.855183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.855212 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.855222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.855235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.855244 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.958405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.958439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.958447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.958460 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:38 crc kubenswrapper[4758]: I1004 10:51:38.958469 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:38Z","lastTransitionTime":"2025-10-04T10:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.060559 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.060639 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.060676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.060694 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.060707 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.164318 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.164393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.164413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.164467 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.164487 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.267301 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.267367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.267385 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.267409 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.267432 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.325559 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.325634 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.325577 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:39 crc kubenswrapper[4758]: E1004 10:51:39.325720 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.325807 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:39 crc kubenswrapper[4758]: E1004 10:51:39.325931 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:39 crc kubenswrapper[4758]: E1004 10:51:39.325983 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:39 crc kubenswrapper[4758]: E1004 10:51:39.326187 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.370240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.370286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.370295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.370312 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.370322 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.472822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.472860 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.472869 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.472884 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.472894 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.575637 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.575702 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.575720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.575743 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.575759 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.678180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.678260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.678282 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.678314 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.678336 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.781953 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.782021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.782038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.782065 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.782084 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.885272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.885338 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.885360 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.885387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.885408 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.988067 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.988161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.988180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.988202 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:39 crc kubenswrapper[4758]: I1004 10:51:39.988219 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:39Z","lastTransitionTime":"2025-10-04T10:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.091470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.091515 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.091578 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.091599 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.091657 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.194094 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.194197 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.194243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.194263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.194277 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.296837 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.296900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.296909 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.296923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.296932 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.400085 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.400169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.400186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.400204 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.400215 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.503628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.503692 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.503710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.503731 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.503748 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.606990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.607178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.607209 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.607302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.607526 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.710787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.710841 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.710857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.710879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.710896 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.814387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.814438 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.814455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.814477 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.814494 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.917013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.917051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.917064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.917078 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:40 crc kubenswrapper[4758]: I1004 10:51:40.917090 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:40Z","lastTransitionTime":"2025-10-04T10:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.020787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.020850 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.020868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.020891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.020911 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.123163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.123240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.123262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.123292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.123312 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.226475 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.226520 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.226536 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.226557 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.226575 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.325869 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.325924 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.325896 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.325878 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.326078 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.326231 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.326293 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.326353 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.328320 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.328350 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.328361 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.328375 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.328429 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.432014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.432130 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.432148 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.432169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.432187 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.534806 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.534932 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.534951 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.534980 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.534998 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.637486 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.637589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.637609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.637634 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.637651 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.740836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.740947 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.740967 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.741006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.741025 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.773383 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.773453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.773474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.773506 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.773530 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.794888 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.798582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.798626 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.798636 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.798651 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.798661 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.817042 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.821661 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.821730 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.821744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.821760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.821811 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.837316 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.841366 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.841611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.841793 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.842020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.842452 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.858675 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.864968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.865259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.865470 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.865961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.866369 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.888376 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:41 crc kubenswrapper[4758]: E1004 10:51:41.888614 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.892040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.892323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.892503 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.892685 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.892813 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.995746 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.995800 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.995819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.995844 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:41 crc kubenswrapper[4758]: I1004 10:51:41.995861 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:41Z","lastTransitionTime":"2025-10-04T10:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.098895 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.098948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.098964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.098985 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.099003 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.202040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.202074 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.202084 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.202112 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.202126 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.304783 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.305052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.305141 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.305234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.305295 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.408285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.408534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.408616 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.408719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.408791 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.510922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.511210 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.511317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.511418 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.511499 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.613822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.613879 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.613896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.613920 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.613937 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.716694 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.716772 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.716795 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.716818 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.716834 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.819214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.819468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.819566 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.819654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.819765 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.922816 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.923140 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.923225 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.923323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:42 crc kubenswrapper[4758]: I1004 10:51:42.923414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:42Z","lastTransitionTime":"2025-10-04T10:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.027549 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.027652 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.027680 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.027754 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.027776 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.131820 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.131912 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.131937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.131971 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.131993 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.236187 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.236242 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.236260 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.236282 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.236300 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.325168 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.325178 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.325324 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:43 crc kubenswrapper[4758]: E1004 10:51:43.325367 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:43 crc kubenswrapper[4758]: E1004 10:51:43.325696 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:43 crc kubenswrapper[4758]: E1004 10:51:43.326014 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.327032 4758 scope.go:117] "RemoveContainer" containerID="49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.328597 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:43 crc kubenswrapper[4758]: E1004 10:51:43.330744 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.338254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.338373 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.338494 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.338586 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.338671 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.344145 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.363535 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.383086 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.400191 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.417721 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.441885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.441946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.441956 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.441988 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.442001 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.452665 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.486974 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.510792 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.527478 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.545385 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.547239 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.547518 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.547758 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.547981 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.548215 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.566357 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.584550 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.602925 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.624027 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.643997 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.651399 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.651447 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.651462 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.651502 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.651517 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.660708 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.680638 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.699659 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.754758 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.754814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.755059 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.755151 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.755169 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.839067 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/2.log" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.843034 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.843675 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.857078 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.858411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.858688 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.858707 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.858729 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.858749 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.877137 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.890409 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.908047 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.919942 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.934591 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.955037 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.961662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.961715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.961732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.961753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.961768 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:43Z","lastTransitionTime":"2025-10-04T10:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:43 crc kubenswrapper[4758]: I1004 10:51:43.997440 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:43Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.013984 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.029198 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.046240 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.063436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.063482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.063492 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.063519 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.063531 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.068635 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.079952 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.092422 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.102798 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.113583 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.123199 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.133230 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.165539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.165582 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.165592 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.165610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.165629 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.268220 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.268277 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.268288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.268303 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.268313 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.371207 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.371272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.371288 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.371312 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.371330 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.473576 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.473627 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.473638 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.473656 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.473668 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.576129 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.576175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.576193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.576209 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.576219 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.679942 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.679990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.680011 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.680037 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.680057 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.782900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.782962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.783013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.783040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.783057 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.848791 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/3.log" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.849990 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/2.log" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.854616 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" exitCode=1 Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.854668 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.854708 4758 scope.go:117] "RemoveContainer" containerID="49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.856147 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 10:51:44 crc kubenswrapper[4758]: E1004 10:51:44.856476 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.879563 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.891078 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.891155 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.891174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.891197 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.891213 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.906548 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.926265 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.941537 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.963095 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.984878 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.995621 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.995852 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.995881 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.995911 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:44 crc kubenswrapper[4758]: I1004 10:51:44.995934 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:44Z","lastTransitionTime":"2025-10-04T10:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.001929 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:44Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.016354 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.037601 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.055417 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.071484 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.085250 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.100505 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.100580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.100607 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.100640 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.100664 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.110433 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.133590 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49eccf198d431d666060b759ce57a1a5910ea6e4c54404e2c877eea04f3c6cc7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:17Z\\\",\\\"message\\\":\\\"rollers per-node LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242275 6357 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-qkzfh\\\\nI1004 10:51:17.242285 6357 services_controller.go:453] Built service openshift-machine-api/machine-api-controllers template LB for network=default: []services.LB{}\\\\nI1004 10:51:17.242298 6357 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:51:17.242301 6357 services_controller.go:454] Service openshift-machine-api/machine-api-controllers for network=default has 3 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1004 10:51:17.242322 6357 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}\\\\nI1004 10:51:17.242342 6357 services_controller.go:360] Finished syncing service network-check-target on namespace openshift-network-diagnostics for network=default : 1.174267ms\\\\nF1004 10:51:17.242350 6357 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initializa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:44Z\\\",\\\"message\\\":\\\"hift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-hbnh9 openshift-dns/node-resolver-qkzfh openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc]\\\\nI1004 10:51:44.435723 6698 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1004 10:51:44.435745 6698 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435758 6698 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435793 6698 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc in node crc\\\\nI1004 10:51:44.435804 6698 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1004 10:51:44.435811 6698 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435832 6698 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 10:51:44.435904 6698 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.147954 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.158418 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.172018 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.189817 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.203673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.203736 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.203753 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.203776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.203793 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.307244 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.307305 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.307323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.307347 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.307368 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.326851 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:45 crc kubenswrapper[4758]: E1004 10:51:45.327036 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.327172 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:45 crc kubenswrapper[4758]: E1004 10:51:45.327313 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.327402 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:45 crc kubenswrapper[4758]: E1004 10:51:45.327482 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.327544 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:45 crc kubenswrapper[4758]: E1004 10:51:45.327622 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.410459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.410560 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.410577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.410601 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.410619 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.513081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.513163 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.513178 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.513196 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.513208 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.616247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.616318 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.616335 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.616357 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.616376 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.719474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.719544 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.719571 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.719605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.719631 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.822052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.822566 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.822874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.823021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.823461 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.861147 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/3.log" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.866979 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 10:51:45 crc kubenswrapper[4758]: E1004 10:51:45.867439 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.889237 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.914431 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.927199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.927234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.927267 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.927289 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.927302 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:45Z","lastTransitionTime":"2025-10-04T10:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.938916 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.958826 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.974534 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:45 crc kubenswrapper[4758]: I1004 10:51:45.986985 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:45Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.007245 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.022479 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.029839 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.029887 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.029904 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.029927 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.029942 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.037367 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.063706 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:44Z\\\",\\\"message\\\":\\\"hift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-hbnh9 openshift-dns/node-resolver-qkzfh openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc]\\\\nI1004 10:51:44.435723 6698 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1004 10:51:44.435745 6698 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435758 6698 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435793 6698 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc in node crc\\\\nI1004 10:51:44.435804 6698 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1004 10:51:44.435811 6698 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435832 6698 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 10:51:44.435904 6698 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.077610 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.094263 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.109243 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.122899 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.132185 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.132214 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.132227 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.132243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.132254 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.137450 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.151516 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.162953 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.175917 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:46Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.234268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.234326 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.234382 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.234406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.234438 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.337988 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.338136 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.338166 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.338199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.338520 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.442235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.442636 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.442843 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.443009 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.443189 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.547134 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.547183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.547199 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.547400 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.547414 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.652603 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.652681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.652704 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.652733 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.652756 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.754792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.754830 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.754838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.754854 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.754871 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.856779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.856839 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.856855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.856877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.856892 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.959682 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.959716 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.959726 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.959740 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:46 crc kubenswrapper[4758]: I1004 10:51:46.959749 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:46Z","lastTransitionTime":"2025-10-04T10:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.062937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.063201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.063368 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.063496 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.063612 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.155647 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.155922 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.155896804 +0000 UTC m=+148.448547693 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.156409 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.156494 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.156588 4758 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.156623 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.156615371 +0000 UTC m=+148.449266260 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.156699 4758 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.156722 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.156716514 +0000 UTC m=+148.449367403 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.166052 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.166074 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.166081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.166093 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.166120 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.258322 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.258420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.258625 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.258652 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.258678 4758 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.258750 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.25872653 +0000 UTC m=+148.551377459 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.259043 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.259067 4758 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.259082 4758 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.259175 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.259159611 +0000 UTC m=+148.551810540 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.268993 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.269223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.269362 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.269521 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.269654 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.324864 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.324928 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.325039 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.324879 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.325143 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.325270 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.325396 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:47 crc kubenswrapper[4758]: E1004 10:51:47.325525 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.373026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.373075 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.373088 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.373132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.373149 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.476424 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.476770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.476896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.477019 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.477207 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.580828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.581131 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.581287 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.581419 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.581553 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.685664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.685728 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.685746 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.685771 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.685790 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.789093 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.789191 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.789212 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.789236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.789254 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.892326 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.892384 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.892401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.892423 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.892441 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.995179 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.995235 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.995251 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.995274 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:47 crc kubenswrapper[4758]: I1004 10:51:47.995293 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:47Z","lastTransitionTime":"2025-10-04T10:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.097594 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.097646 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.097662 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.097683 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.097699 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.199824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.199876 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.199891 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.199913 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.199930 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.302642 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.302991 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.303147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.303467 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.303635 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.406609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.406641 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.406653 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.406670 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.406682 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.510228 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.510293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.510350 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.510380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.510403 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.613510 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.613581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.613604 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.613630 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.613653 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.716866 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.716944 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.716964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.716990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.717007 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.820469 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.820590 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.820609 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.821023 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.821072 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.923404 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.923436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.923444 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.923456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:48 crc kubenswrapper[4758]: I1004 10:51:48.923465 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:48Z","lastTransitionTime":"2025-10-04T10:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.025858 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.025920 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.025937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.025962 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.025982 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.129147 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.129327 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.129432 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.129541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.129637 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.231970 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.232023 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.232041 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.232064 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.232082 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.325759 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.325886 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.325923 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:49 crc kubenswrapper[4758]: E1004 10:51:49.326082 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.326325 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:49 crc kubenswrapper[4758]: E1004 10:51:49.326415 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:49 crc kubenswrapper[4758]: E1004 10:51:49.326530 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:49 crc kubenswrapper[4758]: E1004 10:51:49.326296 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.334557 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.334632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.334657 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.334686 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.334709 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.437589 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.437654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.437681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.437712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.437802 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.541176 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.541248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.541268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.541292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.541313 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.644231 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.644284 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.644302 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.644324 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.644341 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.747683 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.747727 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.747742 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.747761 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.747775 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.851216 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.851259 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.851270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.851286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.851297 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.959664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.959765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.959784 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.959807 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:49 crc kubenswrapper[4758]: I1004 10:51:49.959827 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:49Z","lastTransitionTime":"2025-10-04T10:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.062348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.062411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.062427 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.062452 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.062470 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.165300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.165370 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.165387 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.165413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.165431 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.268585 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.268681 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.268699 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.268724 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.268744 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.371941 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.372006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.372025 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.372050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.372071 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.475231 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.475294 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.475310 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.475336 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.475355 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.577859 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.578354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.578388 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.578421 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.578444 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.682222 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.682282 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.682299 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.682324 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.682341 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.786175 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.786237 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.786254 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.786278 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.786296 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.888555 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.888606 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.888617 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.888634 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.888649 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.991687 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.991732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.991748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.991770 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:50 crc kubenswrapper[4758]: I1004 10:51:50.991787 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:50Z","lastTransitionTime":"2025-10-04T10:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.094969 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.095048 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.095071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.095132 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.095151 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.198333 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.198396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.198413 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.198436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.198452 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.301776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.301827 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.301845 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.301872 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.301894 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.325360 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.325418 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.325470 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.325396 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:51 crc kubenswrapper[4758]: E1004 10:51:51.325603 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:51 crc kubenswrapper[4758]: E1004 10:51:51.326134 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:51 crc kubenswrapper[4758]: E1004 10:51:51.326276 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:51 crc kubenswrapper[4758]: E1004 10:51:51.326419 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.404436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.404498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.404517 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.404541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.404559 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.507897 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.507963 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.507982 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.508006 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.508026 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.611217 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.611275 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.611292 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.611315 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.611332 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.714026 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.714092 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.714151 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.714183 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.714211 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.816787 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.816847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.816871 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.816899 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.816921 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.919457 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.919495 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.919505 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.919523 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:51 crc kubenswrapper[4758]: I1004 10:51:51.919535 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:51Z","lastTransitionTime":"2025-10-04T10:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.021540 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.021593 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.021610 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.021632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.021649 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.117311 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.117374 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.117393 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.117417 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.117437 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: E1004 10:51:52.138841 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.143686 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.143759 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.143782 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.143813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.143833 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: E1004 10:51:52.167138 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.178534 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.178595 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.178612 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.178635 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.178651 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: E1004 10:51:52.198933 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.203386 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.203437 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.203455 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.203479 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.203530 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: E1004 10:51:52.221136 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.225394 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.225429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.225439 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.225454 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.225465 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: E1004 10:51:52.243009 4758 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17ee5017-c6ea-48c4-a1fa-99cd3efae497\\\",\\\"systemUUID\\\":\\\"1c23ca6e-c5ec-490c-92c3-929ef50a0040\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:52 crc kubenswrapper[4758]: E1004 10:51:52.243252 4758 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.245474 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.245515 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.245531 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.245551 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.245568 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.347855 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.347923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.347948 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.347979 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.348003 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.450739 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.450792 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.450804 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.450825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.450837 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.554098 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.554192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.554209 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.554238 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.554256 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.657593 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.657654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.657671 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.657697 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.657714 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.760557 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.760611 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.760627 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.760649 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.760666 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.863270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.863331 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.863348 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.863372 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.863390 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.966547 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.966629 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.966645 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.966741 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:52 crc kubenswrapper[4758]: I1004 10:51:52.966761 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:52Z","lastTransitionTime":"2025-10-04T10:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.069059 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.069158 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.069182 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.069211 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.069232 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.172781 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.172864 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.172888 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.172922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.172947 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.276689 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.276754 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.276772 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.276825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.276843 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.326152 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:53 crc kubenswrapper[4758]: E1004 10:51:53.326266 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.326668 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.326711 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.326732 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:53 crc kubenswrapper[4758]: E1004 10:51:53.326772 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:53 crc kubenswrapper[4758]: E1004 10:51:53.326920 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:53 crc kubenswrapper[4758]: E1004 10:51:53.327029 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.341925 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a31306dd-e7c1-4d5b-9145-9241de708bdb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1bafaa6e9f6e206e3ce4dbf52e01e37589e28d60b64f25db935ca37933b3c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://590da6c2ce4b922bed7342711595f8365b0adc71767460ff3a397784caee2834\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.359763 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b7d30f12a4ef25cc0015957177d45fe2d49d77e329dc584570c0000e197d3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.372678 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-srgqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5adcca6-45b0-46ad-ac95-57cce5121dc9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:59Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lgwzr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:59Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-srgqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.379710 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.380334 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.380577 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.380774 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.380943 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.389998 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qkzfh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d218843-c1d0-4118-a718-98a5503a09a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88cf161420f21835751a7c32597b5f464bdb6dfafbe5ff2d59424ef97f1656ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-49zpc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qkzfh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.412976 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af0a535d-83d3-4669-9245-fa99ddcb6a7d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://178730ec03d614571e04f119934ddbf88fcace2832049410820e6dc57cd06701\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6153652f23b662ffdd096bb6e0d4f63c75896a1921e99e1a72ceda72b71f43ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac7da09ee6060d8553b4c3b7b328e5d22cefe3dfac6446bf3cdf9527da2719b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d4d7243fb68e2896ee4b8e66ad8cbef69c6c6079fd98c055e5d5bcdf99e68f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://741b63776c94efc29e8f3a979db14ff8b6348992de89f8e77c650a1bbce4ca8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://519230fd478a9e0b493bb78e991205232dff3e3f4c89415ff12170ffae44e09f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c25d140dd8bccec81d8eed2de0d4ec4460a64d417085363871f5c1a67ad9b21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-slrzz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-l4q8s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.444751 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"063cec2b-782d-4033-a24c-4129a935cf27\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:44Z\\\",\\\"message\\\":\\\"hift-network-operator/network-operator-58b4c7f79c-55gtf openshift-ovn-kubernetes/ovnkube-node-hbnh9 openshift-dns/node-resolver-qkzfh openshift-kube-scheduler/openshift-kube-scheduler-crc openshift-machine-config-operator/kube-rbac-proxy-crio-crc]\\\\nI1004 10:51:44.435723 6698 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1004 10:51:44.435745 6698 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435758 6698 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435793 6698 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc in node crc\\\\nI1004 10:51:44.435804 6698 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1004 10:51:44.435811 6698 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1004 10:51:44.435832 6698 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 10:51:44.435904 6698 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:51:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5tk2m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hbnh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.463592 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"314ebe86-11e3-424d-b415-613fb8b7005a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6a3598d5f822c0cd2ea6e529f37660794689898ecb4b09af4c2eb8b35b546c75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b28ad3c8ef720e7fb2a1fcf8c70b2813adc644fb312d67346d4557622407e76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490d9343b66263183b0f0d18296e41467a2f593a0ab55cef395bdaef23df17f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d595810d0adbae6976d77a90801e8a237fc4219c7e0e8eb179dd4229082d10ca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.483031 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.484363 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.484631 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.484813 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.485010 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.485227 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.502545 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.520710 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb4133223a3bd38efa047ce9b412930ae6cab188e34d2d876bcfcf0a2f0d53d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.537982 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6f154174beacd4db5395e14ba4862fef5994711eca9b6b22dbb045629342074\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f18372d1aabfa1094875395439c956df73bc4a703dfa12ad29c6d74546340ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.552931 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-znxcj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8a97aad-9d66-4f7e-ad8a-d829942dd4c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebb437a57f381fba258f1aaa139419b4c1bb450d43bdd48551820a02e8700311\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9qrpx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-znxcj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.567508 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4fd2c8d-474b-48ac-a369-ce63fbf61413\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01ad65f62cb99cae9936bae0eea9de9cca83daf747086326b5a3edad62d6555e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zndv9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-d4w2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.585599 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bpvxr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3875fe1d-de8c-4266-bf2b-e07c633b85dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:51:34Z\\\",\\\"message\\\":\\\"2025-10-04T10:50:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634\\\\n2025-10-04T10:50:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9ec19997-965d-4489-ac15-81946c4d5634 to /host/opt/cni/bin/\\\\n2025-10-04T10:50:49Z [verbose] multus-daemon started\\\\n2025-10-04T10:50:49Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:51:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:46Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctvk5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:45Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bpvxr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.588367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.588411 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.588429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.588453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.588472 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.600245 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84cf9e92-5435-48c3-8035-384e9455e6fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:51:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6452ec305107a2da2260902295d7d310501dd35a8e2c5255f5177f5a378c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2085549166f4cf5a59a644a6df9d068f20b7c47d24b5b5f5e3fd0fbf03516bd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:51:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wptfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jx7sf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.613616 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d40e8ff-0bc5-4f02-88e3-026e80284d2f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e44022535a9d399c3566074645999b77b80a271caa36ee46efa35dd94312c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94365e595aa50817a754b6ce0e22117f7964504e5840c062788b63381ca04abf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdd914de02ec6abe4706e6f61a9d9fb55d6e5ac2d6f8a65b28f3c50d488342b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76bb0fa9dfaca2e09a3e31aa9495069e24cf08e47b2f21230735df39f082f1d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1b49f2b9968651a19afcbaa0fd56303a0f53daf46b297735fe52d50881e8bba\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:50:37Z\\\",\\\"message\\\":\\\"W1004 10:50:26.800392 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 10:50:26.800779 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759575026 cert, and key in /tmp/serving-cert-588764432/serving-signer.crt, /tmp/serving-cert-588764432/serving-signer.key\\\\nI1004 10:50:27.040932 1 observer_polling.go:159] Starting file observer\\\\nW1004 10:50:27.049341 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 10:50:27.049515 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 10:50:27.066163 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-588764432/tls.crt::/tmp/serving-cert-588764432/tls.key\\\\\\\"\\\\nF1004 10:50:37.570016 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc00bc93fd5196ce57fa070eb5ade2994f76c2016cac3735185129ee0a1f9132\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d04440d490c6ee7725ff1ec48d4745cb324220dfd3283cbf3c06bd8c2e4a747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:50:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.626314 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40a3d658-7861-4317-8302-6870205a4fb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://209ebef510d6fab99b001c0412d55cf8167cdd74fe84824044e24dc010fc2d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://051086f8057df6643d1bd46aa5a9e81e823f8655bab5fb7bb88eb075401ff804\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5018d24f77d1d16e49ca50d03bd164e7922e97fb5ed956fa95d0cee4ac59747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc060dadd5b477741f2ad0e3be92e153c9ec70f4f62ce37239d6492b54502d21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:50:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:50:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.639965 4758 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:50:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:51:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.690316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.690599 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.690719 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.690838 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.690948 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.793665 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.793722 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.793737 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.793760 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.793776 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.896016 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.896087 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.896133 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.896162 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.896183 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:53Z","lastTransitionTime":"2025-10-04T10:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:53 crc kubenswrapper[4758]: I1004 10:51:53.999673 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.000021 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.000193 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.000339 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.000592 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.104278 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.104628 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.104789 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.104951 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.105146 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.208343 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.208406 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.208428 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.208453 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.208470 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.312287 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.312580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.312712 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.312882 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.313000 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.345971 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.416655 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.416715 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.416726 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.416744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.416758 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.519726 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.519769 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.519776 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.519790 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.519799 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.623530 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.623580 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.623596 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.623624 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.623644 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.726608 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.726647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.726658 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.726675 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.726687 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.828599 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.828622 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.828632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.828646 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.828654 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.931396 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.931444 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.931459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.931481 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:54 crc kubenswrapper[4758]: I1004 10:51:54.931497 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:54Z","lastTransitionTime":"2025-10-04T10:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.034696 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.034819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.034846 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.034871 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.034890 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.138701 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.138790 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.138812 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.138840 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.138875 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.242274 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.242328 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.242344 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.242367 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.242383 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.325887 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.325907 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.326170 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.326202 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:55 crc kubenswrapper[4758]: E1004 10:51:55.326795 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:55 crc kubenswrapper[4758]: E1004 10:51:55.326908 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:55 crc kubenswrapper[4758]: E1004 10:51:55.326984 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:55 crc kubenswrapper[4758]: E1004 10:51:55.327075 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.344764 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.344808 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.344825 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.344847 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.344866 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.447397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.447460 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.447483 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.447507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.447527 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.550240 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.550286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.550298 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.550316 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.550329 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.652713 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.652768 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.652783 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.652802 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.652814 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.755819 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.755929 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.756272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.756323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.756347 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.859201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.859272 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.859295 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.859327 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.859351 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.962409 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.962456 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.962468 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.962484 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:55 crc kubenswrapper[4758]: I1004 10:51:55.962495 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:55Z","lastTransitionTime":"2025-10-04T10:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.064937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.064996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.065013 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.065038 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.065055 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.167602 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.167659 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.167700 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.167723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.167739 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.270053 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.270089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.270109 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.270123 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.270132 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.373642 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.373676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.373684 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.373698 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.373707 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.476828 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.476918 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.476940 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.476964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.476982 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.579223 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.579291 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.579308 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.579356 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.579374 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.686775 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.686834 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.686851 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.686874 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.686892 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.789944 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.790005 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.790022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.790046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.790065 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.893400 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.893463 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.893482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.893507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.893524 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.995922 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.995961 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.995974 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.995991 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:56 crc kubenswrapper[4758]: I1004 10:51:56.996003 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:56Z","lastTransitionTime":"2025-10-04T10:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.100050 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.100169 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.100189 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.100248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.100267 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.202938 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.202972 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.202981 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.202996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.203007 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.307022 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.307186 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.307210 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.307285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.307539 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.325462 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:57 crc kubenswrapper[4758]: E1004 10:51:57.325619 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.325873 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:57 crc kubenswrapper[4758]: E1004 10:51:57.326083 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.326363 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:57 crc kubenswrapper[4758]: E1004 10:51:57.326463 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.326705 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:57 crc kubenswrapper[4758]: E1004 10:51:57.326939 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.410555 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.410632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.410688 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.410720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.410740 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.514180 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.514263 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.514286 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.514313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.514331 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.617937 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.618020 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.618056 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.618087 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.618177 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.721654 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.722043 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.722248 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.722436 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.722596 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.827198 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.827270 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.827313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.827341 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.827360 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.930632 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.930668 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.930676 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.930688 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:57 crc kubenswrapper[4758]: I1004 10:51:57.930697 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:57Z","lastTransitionTime":"2025-10-04T10:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.033886 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.033954 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.033972 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.033996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.034016 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.136621 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.136720 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.136732 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.136748 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.136761 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.238801 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.238877 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.238900 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.238934 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.238957 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.326873 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 10:51:58 crc kubenswrapper[4758]: E1004 10:51:58.327156 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.342137 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.342206 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.342224 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.342247 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.342264 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.445460 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.445541 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.445567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.445597 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.445620 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.549023 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.549071 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.549089 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.549149 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.549167 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.653909 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.653968 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.653986 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.654012 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.654032 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.757273 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.757340 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.757356 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.757380 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.757398 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.861023 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.861081 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.861124 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.861148 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.861165 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.963983 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.964047 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.964063 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.964088 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:58 crc kubenswrapper[4758]: I1004 10:51:58.964154 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:58Z","lastTransitionTime":"2025-10-04T10:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.067354 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.067418 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.067435 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.067459 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.067476 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.170122 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.170192 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.170210 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.170234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.170262 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.272581 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.272647 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.272664 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.272691 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.272709 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.325219 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.325241 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.325225 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:51:59 crc kubenswrapper[4758]: E1004 10:51:59.325366 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:51:59 crc kubenswrapper[4758]: E1004 10:51:59.325489 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:51:59 crc kubenswrapper[4758]: E1004 10:51:59.325695 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.325800 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:51:59 crc kubenswrapper[4758]: E1004 10:51:59.325989 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.375203 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.375250 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.375265 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.375285 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.375303 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.478353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.478422 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.478445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.478477 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.478499 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.581972 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.582051 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.582069 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.582143 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.582162 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.685234 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.685293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.685313 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.685339 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.685358 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.788371 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.788711 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.788868 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.789014 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.789187 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.892642 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.893037 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.893257 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.893476 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.893618 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.997334 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.997397 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.997412 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.997435 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:51:59 crc kubenswrapper[4758]: I1004 10:51:59.997452 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:51:59Z","lastTransitionTime":"2025-10-04T10:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.101201 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.101300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.101317 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.101342 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.101358 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.204523 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.204584 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.204603 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.204627 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.204646 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.307750 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.307836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.307865 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.307896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.307921 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.411507 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.411995 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.412243 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.412479 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.412750 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.515885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.516184 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.516299 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.516401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.516498 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.623744 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.623814 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.623832 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.623857 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.623885 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.727765 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.727824 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.727840 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.727866 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.727904 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.830405 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.830498 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.830513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.830533 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.830547 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.932836 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.932885 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.932898 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.932917 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:00 crc kubenswrapper[4758]: I1004 10:52:00.932931 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:00Z","lastTransitionTime":"2025-10-04T10:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.036046 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.036466 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.036688 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.036903 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.037153 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.139896 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.139946 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.139996 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.140016 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.140030 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.242835 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.242909 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.242935 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.242964 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.242989 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.325624 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.325659 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:01 crc kubenswrapper[4758]: E1004 10:52:01.326512 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.326576 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:01 crc kubenswrapper[4758]: E1004 10:52:01.326716 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.327194 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:01 crc kubenswrapper[4758]: E1004 10:52:01.327358 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:01 crc kubenswrapper[4758]: E1004 10:52:01.327431 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.345950 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.346219 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.346353 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.346482 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.346606 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.450208 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.450268 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.450293 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.450323 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.450346 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.553822 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.554262 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.554502 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.554954 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.555360 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.658300 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.658376 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.658401 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.658429 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.658450 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.761716 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.761779 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.761798 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.761820 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.761837 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.864723 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.865213 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.865485 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.865722 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.865920 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.968990 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.969090 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.969174 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.969197 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:01 crc kubenswrapper[4758]: I1004 10:52:01.969253 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:01Z","lastTransitionTime":"2025-10-04T10:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.072445 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.072500 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.072516 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.072539 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.072556 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:02Z","lastTransitionTime":"2025-10-04T10:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.175513 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.175567 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.175583 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.175605 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.175621 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:02Z","lastTransitionTime":"2025-10-04T10:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.278161 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.278236 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.278252 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.278274 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.278292 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:02Z","lastTransitionTime":"2025-10-04T10:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.381040 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.381142 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.381164 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.381191 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.381209 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:02Z","lastTransitionTime":"2025-10-04T10:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.484823 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.484889 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.484906 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.484930 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.484951 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:02Z","lastTransitionTime":"2025-10-04T10:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.500923 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.500975 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.500994 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.501015 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.501031 4758 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:52:02Z","lastTransitionTime":"2025-10-04T10:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.552430 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz"] Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.553064 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.555374 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.555639 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.555771 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.555830 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.582511 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=32.582487962 podStartE2EDuration="32.582487962s" podCreationTimestamp="2025-10-04 10:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.582310828 +0000 UTC m=+99.874961757" watchObservedRunningTime="2025-10-04 10:52:02.582487962 +0000 UTC m=+99.875138881" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.631772 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/16605561-a193-46b2-9f38-dd37d289895b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.631845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16605561-a193-46b2-9f38-dd37d289895b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.631884 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/16605561-a193-46b2-9f38-dd37d289895b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.631933 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16605561-a193-46b2-9f38-dd37d289895b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.631984 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16605561-a193-46b2-9f38-dd37d289895b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.661778 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qkzfh" podStartSLOduration=77.661762764 podStartE2EDuration="1m17.661762764s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.64368158 +0000 UTC m=+99.936332489" watchObservedRunningTime="2025-10-04 10:52:02.661762764 +0000 UTC m=+99.954413653" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.688087 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-l4q8s" podStartSLOduration=77.688067685 podStartE2EDuration="1m17.688067685s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.662351748 +0000 UTC m=+99.955002647" watchObservedRunningTime="2025-10-04 10:52:02.688067685 +0000 UTC m=+99.980718584" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.717025 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.716998238 podStartE2EDuration="44.716998238s" podCreationTimestamp="2025-10-04 10:51:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.704043318 +0000 UTC m=+99.996694247" watchObservedRunningTime="2025-10-04 10:52:02.716998238 +0000 UTC m=+100.009649167" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.727013 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podStartSLOduration=77.726994538 podStartE2EDuration="1m17.726994538s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.726854875 +0000 UTC m=+100.019505804" watchObservedRunningTime="2025-10-04 10:52:02.726994538 +0000 UTC m=+100.019645427" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.732615 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/16605561-a193-46b2-9f38-dd37d289895b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.732861 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16605561-a193-46b2-9f38-dd37d289895b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.732974 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16605561-a193-46b2-9f38-dd37d289895b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.732712 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/16605561-a193-46b2-9f38-dd37d289895b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.733241 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16605561-a193-46b2-9f38-dd37d289895b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.733403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/16605561-a193-46b2-9f38-dd37d289895b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.733525 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/16605561-a193-46b2-9f38-dd37d289895b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.733632 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/16605561-a193-46b2-9f38-dd37d289895b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.753757 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16605561-a193-46b2-9f38-dd37d289895b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.757019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16605561-a193-46b2-9f38-dd37d289895b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-bf9bz\" (UID: \"16605561-a193-46b2-9f38-dd37d289895b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.798780 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=8.7987628 podStartE2EDuration="8.7987628s" podCreationTimestamp="2025-10-04 10:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.797777916 +0000 UTC m=+100.090428805" watchObservedRunningTime="2025-10-04 10:52:02.7987628 +0000 UTC m=+100.091413689" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.799054 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-znxcj" podStartSLOduration=77.799048146 podStartE2EDuration="1m17.799048146s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.769816925 +0000 UTC m=+100.062467814" watchObservedRunningTime="2025-10-04 10:52:02.799048146 +0000 UTC m=+100.091699035" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.837931 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jx7sf" podStartSLOduration=77.837917179 podStartE2EDuration="1m17.837917179s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.837787826 +0000 UTC m=+100.130438715" watchObservedRunningTime="2025-10-04 10:52:02.837917179 +0000 UTC m=+100.130568068" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.838315 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-bpvxr" podStartSLOduration=77.838310298 podStartE2EDuration="1m17.838310298s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.826578817 +0000 UTC m=+100.119229706" watchObservedRunningTime="2025-10-04 10:52:02.838310298 +0000 UTC m=+100.130961177" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.870882 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.870865279 podStartE2EDuration="1m19.870865279s" podCreationTimestamp="2025-10-04 10:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.855601673 +0000 UTC m=+100.148252562" watchObservedRunningTime="2025-10-04 10:52:02.870865279 +0000 UTC m=+100.163516168" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.870986 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.870983072 podStartE2EDuration="1m19.870983072s" podCreationTimestamp="2025-10-04 10:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:02.870765597 +0000 UTC m=+100.163416486" watchObservedRunningTime="2025-10-04 10:52:02.870983072 +0000 UTC m=+100.163633961" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.875819 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" Oct 04 10:52:02 crc kubenswrapper[4758]: I1004 10:52:02.930449 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" event={"ID":"16605561-a193-46b2-9f38-dd37d289895b","Type":"ContainerStarted","Data":"27636cdfe8ebee8415118c56502a67c3cdd4d22dbdba8e8c4ed805d412d59d70"} Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.324893 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.324904 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.324956 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.325344 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:03 crc kubenswrapper[4758]: E1004 10:52:03.326135 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:03 crc kubenswrapper[4758]: E1004 10:52:03.326313 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:03 crc kubenswrapper[4758]: E1004 10:52:03.326445 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:03 crc kubenswrapper[4758]: E1004 10:52:03.326542 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.642014 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:03 crc kubenswrapper[4758]: E1004 10:52:03.642364 4758 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:52:03 crc kubenswrapper[4758]: E1004 10:52:03.642480 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs podName:e5adcca6-45b0-46ad-ac95-57cce5121dc9 nodeName:}" failed. No retries permitted until 2025-10-04 10:53:07.642449926 +0000 UTC m=+164.935100855 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs") pod "network-metrics-daemon-srgqp" (UID: "e5adcca6-45b0-46ad-ac95-57cce5121dc9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.936507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" event={"ID":"16605561-a193-46b2-9f38-dd37d289895b","Type":"ContainerStarted","Data":"9f764bfc96bd432ea946de09f2a45e14bde89bdb5e2bb1e59c7c3a0a06dfd85c"} Oct 04 10:52:03 crc kubenswrapper[4758]: I1004 10:52:03.957558 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-bf9bz" podStartSLOduration=78.957534493 podStartE2EDuration="1m18.957534493s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:03.955922105 +0000 UTC m=+101.248573024" watchObservedRunningTime="2025-10-04 10:52:03.957534493 +0000 UTC m=+101.250185412" Oct 04 10:52:05 crc kubenswrapper[4758]: I1004 10:52:05.324816 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:05 crc kubenswrapper[4758]: E1004 10:52:05.324971 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:05 crc kubenswrapper[4758]: I1004 10:52:05.324970 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:05 crc kubenswrapper[4758]: E1004 10:52:05.325059 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:05 crc kubenswrapper[4758]: I1004 10:52:05.325002 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:05 crc kubenswrapper[4758]: E1004 10:52:05.325163 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:05 crc kubenswrapper[4758]: I1004 10:52:05.326763 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:05 crc kubenswrapper[4758]: E1004 10:52:05.327915 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:07 crc kubenswrapper[4758]: I1004 10:52:07.325404 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:07 crc kubenswrapper[4758]: I1004 10:52:07.325412 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:07 crc kubenswrapper[4758]: I1004 10:52:07.325507 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:07 crc kubenswrapper[4758]: E1004 10:52:07.325657 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:07 crc kubenswrapper[4758]: I1004 10:52:07.325870 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:07 crc kubenswrapper[4758]: E1004 10:52:07.325953 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:07 crc kubenswrapper[4758]: E1004 10:52:07.326142 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:07 crc kubenswrapper[4758]: E1004 10:52:07.326373 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:09 crc kubenswrapper[4758]: I1004 10:52:09.324791 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:09 crc kubenswrapper[4758]: I1004 10:52:09.324869 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:09 crc kubenswrapper[4758]: I1004 10:52:09.325971 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:09 crc kubenswrapper[4758]: E1004 10:52:09.326265 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:09 crc kubenswrapper[4758]: I1004 10:52:09.326337 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:09 crc kubenswrapper[4758]: E1004 10:52:09.326499 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:09 crc kubenswrapper[4758]: E1004 10:52:09.326705 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:09 crc kubenswrapper[4758]: E1004 10:52:09.326847 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:11 crc kubenswrapper[4758]: I1004 10:52:11.325647 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:11 crc kubenswrapper[4758]: E1004 10:52:11.325899 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:11 crc kubenswrapper[4758]: I1004 10:52:11.325918 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:11 crc kubenswrapper[4758]: I1004 10:52:11.325942 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:11 crc kubenswrapper[4758]: I1004 10:52:11.325993 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:11 crc kubenswrapper[4758]: I1004 10:52:11.326589 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 10:52:11 crc kubenswrapper[4758]: E1004 10:52:11.326759 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:52:11 crc kubenswrapper[4758]: E1004 10:52:11.326856 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:11 crc kubenswrapper[4758]: E1004 10:52:11.326920 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:11 crc kubenswrapper[4758]: E1004 10:52:11.326969 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:13 crc kubenswrapper[4758]: I1004 10:52:13.326694 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:13 crc kubenswrapper[4758]: I1004 10:52:13.327631 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:13 crc kubenswrapper[4758]: I1004 10:52:13.327904 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:13 crc kubenswrapper[4758]: E1004 10:52:13.328897 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:13 crc kubenswrapper[4758]: I1004 10:52:13.328943 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:13 crc kubenswrapper[4758]: E1004 10:52:13.329164 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:13 crc kubenswrapper[4758]: E1004 10:52:13.329377 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:13 crc kubenswrapper[4758]: E1004 10:52:13.329534 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:15 crc kubenswrapper[4758]: I1004 10:52:15.325224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:15 crc kubenswrapper[4758]: I1004 10:52:15.325185 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:15 crc kubenswrapper[4758]: I1004 10:52:15.325185 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:15 crc kubenswrapper[4758]: E1004 10:52:15.325431 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:15 crc kubenswrapper[4758]: I1004 10:52:15.325534 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:15 crc kubenswrapper[4758]: E1004 10:52:15.325653 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:15 crc kubenswrapper[4758]: E1004 10:52:15.325934 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:15 crc kubenswrapper[4758]: E1004 10:52:15.326036 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:17 crc kubenswrapper[4758]: I1004 10:52:17.326182 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:17 crc kubenswrapper[4758]: I1004 10:52:17.326349 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:17 crc kubenswrapper[4758]: I1004 10:52:17.326435 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:17 crc kubenswrapper[4758]: E1004 10:52:17.326533 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:17 crc kubenswrapper[4758]: I1004 10:52:17.326550 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:17 crc kubenswrapper[4758]: E1004 10:52:17.326680 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:17 crc kubenswrapper[4758]: E1004 10:52:17.326860 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:17 crc kubenswrapper[4758]: E1004 10:52:17.326954 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:19 crc kubenswrapper[4758]: I1004 10:52:19.325784 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:19 crc kubenswrapper[4758]: I1004 10:52:19.325784 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:19 crc kubenswrapper[4758]: I1004 10:52:19.325858 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:19 crc kubenswrapper[4758]: I1004 10:52:19.325917 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:19 crc kubenswrapper[4758]: E1004 10:52:19.328000 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:19 crc kubenswrapper[4758]: E1004 10:52:19.328252 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:19 crc kubenswrapper[4758]: E1004 10:52:19.328461 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:19 crc kubenswrapper[4758]: E1004 10:52:19.328706 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:20 crc kubenswrapper[4758]: I1004 10:52:20.994075 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/1.log" Oct 04 10:52:20 crc kubenswrapper[4758]: I1004 10:52:20.995491 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/0.log" Oct 04 10:52:20 crc kubenswrapper[4758]: I1004 10:52:20.995651 4758 generic.go:334] "Generic (PLEG): container finished" podID="3875fe1d-de8c-4266-bf2b-e07c633b85dc" containerID="b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e" exitCode=1 Oct 04 10:52:20 crc kubenswrapper[4758]: I1004 10:52:20.995769 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerDied","Data":"b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e"} Oct 04 10:52:20 crc kubenswrapper[4758]: I1004 10:52:20.995848 4758 scope.go:117] "RemoveContainer" containerID="13b6cb2f4d7e54b3d18603df1cf7dd5ca3e23798d3014da27aaa8037dc66f4f1" Oct 04 10:52:20 crc kubenswrapper[4758]: I1004 10:52:20.996639 4758 scope.go:117] "RemoveContainer" containerID="b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e" Oct 04 10:52:20 crc kubenswrapper[4758]: E1004 10:52:20.997153 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-bpvxr_openshift-multus(3875fe1d-de8c-4266-bf2b-e07c633b85dc)\"" pod="openshift-multus/multus-bpvxr" podUID="3875fe1d-de8c-4266-bf2b-e07c633b85dc" Oct 04 10:52:21 crc kubenswrapper[4758]: I1004 10:52:21.325516 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:21 crc kubenswrapper[4758]: E1004 10:52:21.325632 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:21 crc kubenswrapper[4758]: I1004 10:52:21.325734 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:21 crc kubenswrapper[4758]: E1004 10:52:21.325804 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:21 crc kubenswrapper[4758]: I1004 10:52:21.326053 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:21 crc kubenswrapper[4758]: I1004 10:52:21.326136 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:21 crc kubenswrapper[4758]: E1004 10:52:21.326182 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:21 crc kubenswrapper[4758]: E1004 10:52:21.326271 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:22 crc kubenswrapper[4758]: I1004 10:52:22.001025 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/1.log" Oct 04 10:52:23 crc kubenswrapper[4758]: I1004 10:52:23.324971 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:23 crc kubenswrapper[4758]: I1004 10:52:23.325005 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:23 crc kubenswrapper[4758]: I1004 10:52:23.325038 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:23 crc kubenswrapper[4758]: I1004 10:52:23.324971 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.327571 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.327779 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.327907 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.328010 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:23 crc kubenswrapper[4758]: I1004 10:52:23.329239 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.329556 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hbnh9_openshift-ovn-kubernetes(063cec2b-782d-4033-a24c-4129a935cf27)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.378651 4758 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 10:52:23 crc kubenswrapper[4758]: E1004 10:52:23.447917 4758 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 10:52:25 crc kubenswrapper[4758]: I1004 10:52:25.325654 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:25 crc kubenswrapper[4758]: I1004 10:52:25.325757 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:25 crc kubenswrapper[4758]: E1004 10:52:25.325888 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:25 crc kubenswrapper[4758]: I1004 10:52:25.326151 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:25 crc kubenswrapper[4758]: I1004 10:52:25.326206 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:25 crc kubenswrapper[4758]: E1004 10:52:25.326276 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:25 crc kubenswrapper[4758]: E1004 10:52:25.326382 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:25 crc kubenswrapper[4758]: E1004 10:52:25.326491 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:27 crc kubenswrapper[4758]: I1004 10:52:27.325457 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:27 crc kubenswrapper[4758]: I1004 10:52:27.325503 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:27 crc kubenswrapper[4758]: E1004 10:52:27.325632 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:27 crc kubenswrapper[4758]: I1004 10:52:27.325652 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:27 crc kubenswrapper[4758]: E1004 10:52:27.325774 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:27 crc kubenswrapper[4758]: I1004 10:52:27.326187 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:27 crc kubenswrapper[4758]: E1004 10:52:27.326352 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:27 crc kubenswrapper[4758]: E1004 10:52:27.326749 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:28 crc kubenswrapper[4758]: E1004 10:52:28.449092 4758 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 10:52:29 crc kubenswrapper[4758]: I1004 10:52:29.325710 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:29 crc kubenswrapper[4758]: I1004 10:52:29.325721 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:29 crc kubenswrapper[4758]: I1004 10:52:29.325746 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:29 crc kubenswrapper[4758]: E1004 10:52:29.326509 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:29 crc kubenswrapper[4758]: I1004 10:52:29.325819 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:29 crc kubenswrapper[4758]: E1004 10:52:29.326648 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:29 crc kubenswrapper[4758]: E1004 10:52:29.326771 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:29 crc kubenswrapper[4758]: E1004 10:52:29.327082 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:31 crc kubenswrapper[4758]: I1004 10:52:31.325044 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:31 crc kubenswrapper[4758]: E1004 10:52:31.325276 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:31 crc kubenswrapper[4758]: I1004 10:52:31.326157 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:31 crc kubenswrapper[4758]: I1004 10:52:31.326406 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:31 crc kubenswrapper[4758]: I1004 10:52:31.326516 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:31 crc kubenswrapper[4758]: E1004 10:52:31.326788 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:31 crc kubenswrapper[4758]: E1004 10:52:31.327348 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:31 crc kubenswrapper[4758]: E1004 10:52:31.327919 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:32 crc kubenswrapper[4758]: I1004 10:52:32.326805 4758 scope.go:117] "RemoveContainer" containerID="b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e" Oct 04 10:52:33 crc kubenswrapper[4758]: I1004 10:52:33.051023 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/1.log" Oct 04 10:52:33 crc kubenswrapper[4758]: I1004 10:52:33.051076 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerStarted","Data":"1570a60d9f80108e4ebf690d60a5ecffd03622708015353cc0f5247388c750f8"} Oct 04 10:52:33 crc kubenswrapper[4758]: I1004 10:52:33.325339 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:33 crc kubenswrapper[4758]: I1004 10:52:33.325673 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:33 crc kubenswrapper[4758]: E1004 10:52:33.326307 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:33 crc kubenswrapper[4758]: I1004 10:52:33.326322 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:33 crc kubenswrapper[4758]: I1004 10:52:33.326493 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:33 crc kubenswrapper[4758]: E1004 10:52:33.326567 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:33 crc kubenswrapper[4758]: E1004 10:52:33.326672 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:33 crc kubenswrapper[4758]: E1004 10:52:33.326770 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:33 crc kubenswrapper[4758]: E1004 10:52:33.449700 4758 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 10:52:35 crc kubenswrapper[4758]: I1004 10:52:35.325812 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:35 crc kubenswrapper[4758]: I1004 10:52:35.325926 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:35 crc kubenswrapper[4758]: E1004 10:52:35.326003 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:35 crc kubenswrapper[4758]: I1004 10:52:35.326159 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:35 crc kubenswrapper[4758]: E1004 10:52:35.326211 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:35 crc kubenswrapper[4758]: E1004 10:52:35.326467 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:35 crc kubenswrapper[4758]: I1004 10:52:35.326502 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:35 crc kubenswrapper[4758]: E1004 10:52:35.326661 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:36 crc kubenswrapper[4758]: I1004 10:52:36.326001 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.067890 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/3.log" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.069931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerStarted","Data":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.070915 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.109816 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podStartSLOduration=112.109797776 podStartE2EDuration="1m52.109797776s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:37.108326049 +0000 UTC m=+134.400976938" watchObservedRunningTime="2025-10-04 10:52:37.109797776 +0000 UTC m=+134.402448665" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.258835 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-srgqp"] Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.258929 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:37 crc kubenswrapper[4758]: E1004 10:52:37.259005 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.324941 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.324962 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:37 crc kubenswrapper[4758]: E1004 10:52:37.325052 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:37 crc kubenswrapper[4758]: I1004 10:52:37.325089 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:37 crc kubenswrapper[4758]: E1004 10:52:37.325293 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:37 crc kubenswrapper[4758]: E1004 10:52:37.325417 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:38 crc kubenswrapper[4758]: E1004 10:52:38.452026 4758 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 10:52:39 crc kubenswrapper[4758]: I1004 10:52:39.325057 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:39 crc kubenswrapper[4758]: I1004 10:52:39.325180 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:39 crc kubenswrapper[4758]: E1004 10:52:39.325281 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:39 crc kubenswrapper[4758]: I1004 10:52:39.325079 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:39 crc kubenswrapper[4758]: E1004 10:52:39.325573 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:39 crc kubenswrapper[4758]: I1004 10:52:39.325929 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:39 crc kubenswrapper[4758]: E1004 10:52:39.325933 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:39 crc kubenswrapper[4758]: E1004 10:52:39.326256 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:41 crc kubenswrapper[4758]: I1004 10:52:41.325178 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:41 crc kubenswrapper[4758]: I1004 10:52:41.325189 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:41 crc kubenswrapper[4758]: I1004 10:52:41.325192 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:41 crc kubenswrapper[4758]: E1004 10:52:41.325490 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:41 crc kubenswrapper[4758]: E1004 10:52:41.325365 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:41 crc kubenswrapper[4758]: I1004 10:52:41.325555 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:41 crc kubenswrapper[4758]: E1004 10:52:41.325617 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:41 crc kubenswrapper[4758]: E1004 10:52:41.325676 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.325092 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.325166 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.325229 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.328057 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:43 crc kubenswrapper[4758]: E1004 10:52:43.328087 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:52:43 crc kubenswrapper[4758]: E1004 10:52:43.328238 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:52:43 crc kubenswrapper[4758]: E1004 10:52:43.328346 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:52:43 crc kubenswrapper[4758]: E1004 10:52:43.328453 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-srgqp" podUID="e5adcca6-45b0-46ad-ac95-57cce5121dc9" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.594093 4758 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.648752 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8zvst"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.649665 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.651678 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.652473 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.660629 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.661353 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.661402 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.673864 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.678440 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.678911 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.679220 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.679674 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.679884 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.680020 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.679683 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.680696 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.681894 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4z96l"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.706054 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.706951 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pv2mm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.707416 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.707456 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.707785 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gjvgj"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.708046 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.708677 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.710294 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.711058 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.711382 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.719797 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sdqlk"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.720261 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8zmq8"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.720480 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.720817 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.720479 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.721130 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.721377 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4h92h"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.721876 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722013 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722081 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722317 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722510 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722791 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722921 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722922 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.723648 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.722806 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.723211 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.723262 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.723269 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.723308 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.724350 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.728224 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dwm2t"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.733215 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.733654 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.727670 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.734084 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.734267 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.734350 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.734620 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.735713 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.727923 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.736265 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.736362 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.736423 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5ff4f"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.727978 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.736582 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.728044 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.728199 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.728413 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.736839 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.736912 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.728487 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.737302 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.737737 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.745237 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.746915 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.747309 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.747814 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.747954 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.748171 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.748372 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.748533 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.752324 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.759431 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8zvst"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.759462 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.760364 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.764761 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.767622 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.768062 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.769088 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.769656 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.770066 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.775630 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.776399 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.796197 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.796822 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.797009 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.797392 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.801382 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.810306 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.810832 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.811131 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gr62g"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.811424 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.811760 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.812092 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.812243 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.812371 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.812509 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817725 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-client-ca\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817763 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-etcd-serving-ca\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817783 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221c0f36-52eb-4c35-b5b1-7c59b72caff5-serving-cert\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817800 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/957b0629-df4e-473b-9c09-eb9430312c69-node-pullsecrets\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817828 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-image-import-ca\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817856 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817871 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817896 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-serving-cert\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817918 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817935 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957b0629-df4e-473b-9c09-eb9430312c69-audit-dir\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817950 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817963 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/221c0f36-52eb-4c35-b5b1-7c59b72caff5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817979 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxghh\" (UniqueName: \"kubernetes.io/projected/1b358a58-bcec-4a77-9031-64915fa2246d-kube-api-access-sxghh\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.817997 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f8d7281-c6e3-4614-978e-ae5d4c170a85-serving-cert\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818011 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1b358a58-bcec-4a77-9031-64915fa2246d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818026 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkh7s\" (UniqueName: \"kubernetes.io/projected/9af337b4-75ac-4e43-8131-73d96f883e16-kube-api-access-gkh7s\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818042 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/49a1ec68-f811-41d8-a01a-1c692341a7ba-auth-proxy-config\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818056 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/62cf7efa-32a9-4564-875e-b7b6b619805f-images\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818070 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-etcd-client\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818083 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9w6m\" (UniqueName: \"kubernetes.io/projected/1f8d7281-c6e3-4614-978e-ae5d4c170a85-kube-api-access-t9w6m\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818114 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b358a58-bcec-4a77-9031-64915fa2246d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818131 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-policies\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818149 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818163 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-encryption-config\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818177 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f8d7281-c6e3-4614-978e-ae5d4c170a85-config\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818207 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r6jb\" (UniqueName: \"kubernetes.io/projected/62cf7efa-32a9-4564-875e-b7b6b619805f-kube-api-access-5r6jb\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-config\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sdvc\" (UniqueName: \"kubernetes.io/projected/49a1ec68-f811-41d8-a01a-1c692341a7ba-kube-api-access-2sdvc\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818270 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkxhg\" (UniqueName: \"kubernetes.io/projected/4c65edd1-18dd-4b95-81c5-c0a4e5d6013d-kube-api-access-lkxhg\") pod \"cluster-samples-operator-665b6dd947-4wrsp\" (UID: \"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818288 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b358a58-bcec-4a77-9031-64915fa2246d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818303 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818317 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-config\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818331 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818346 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-serving-cert\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818359 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-969r6\" (UniqueName: \"kubernetes.io/projected/83a9a944-0512-4919-b44f-6109a6aa4aa6-kube-api-access-969r6\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818373 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c65edd1-18dd-4b95-81c5-c0a4e5d6013d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4wrsp\" (UID: \"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818387 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-config\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818405 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a9a944-0512-4919-b44f-6109a6aa4aa6-serving-cert\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818420 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49a1ec68-f811-41d8-a01a-1c692341a7ba-config\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818437 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818454 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm2jz\" (UniqueName: \"kubernetes.io/projected/221c0f36-52eb-4c35-b5b1-7c59b72caff5-kube-api-access-cm2jz\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818469 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9af337b4-75ac-4e43-8131-73d96f883e16-serving-cert\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818484 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-client-ca\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818499 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818516 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx6cf\" (UniqueName: \"kubernetes.io/projected/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-kube-api-access-vx6cf\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818531 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8stl\" (UniqueName: \"kubernetes.io/projected/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-kube-api-access-b8stl\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818545 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/62cf7efa-32a9-4564-875e-b7b6b619805f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818558 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6v7\" (UniqueName: \"kubernetes.io/projected/957b0629-df4e-473b-9c09-eb9430312c69-kube-api-access-qr6v7\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818573 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-etcd-client\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818587 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818600 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-audit\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818614 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/49a1ec68-f811-41d8-a01a-1c692341a7ba-machine-approver-tls\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818629 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-dir\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818652 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62cf7efa-32a9-4564-875e-b7b6b619805f-config\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818666 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-serving-cert\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818680 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjngr\" (UniqueName: \"kubernetes.io/projected/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-kube-api-access-zjngr\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818695 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-encryption-config\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818709 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scqjm\" (UniqueName: \"kubernetes.io/projected/dfcb1f56-7b71-4cf6-a92d-8f174194e697-kube-api-access-scqjm\") pod \"downloads-7954f5f757-dwm2t\" (UID: \"dfcb1f56-7b71-4cf6-a92d-8f174194e697\") " pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818725 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-audit-policies\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818739 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818753 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818774 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhbk8\" (UniqueName: \"kubernetes.io/projected/907568c8-1873-41b3-8a58-3651ee7e7922-kube-api-access-hhbk8\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818789 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f8d7281-c6e3-4614-978e-ae5d4c170a85-trusted-ca\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818802 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-service-ca-bundle\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818817 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818832 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818846 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818876 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/907568c8-1873-41b3-8a58-3651ee7e7922-audit-dir\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.818910 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-config\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.819773 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.819903 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.819971 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.820070 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.820151 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.820212 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.820305 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.820648 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.820730 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.821021 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.822618 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.822738 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824284 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-hcsr7"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824508 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824656 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824688 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824757 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824894 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l6frd"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825189 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825391 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824903 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825596 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824915 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825724 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.824972 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825027 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825806 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825057 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825903 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.825697 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826014 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826260 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826496 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826606 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826611 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826686 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826824 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.826850 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.827250 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.829421 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.834323 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.839286 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.839401 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.839693 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.840021 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.840458 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.841466 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.841630 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.842353 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.842686 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.846368 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.848573 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.848731 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s4v4x"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.849373 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.849664 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.850530 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.866199 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.868603 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.869237 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.869943 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.868653 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-pqnhz"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.873998 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.874124 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.874296 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.874397 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.876533 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.877994 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.879189 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.887159 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jf9xm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.890949 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.892258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.893915 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.895083 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.900448 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.900647 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pv2mm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.901007 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.901271 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.901452 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t9c2p"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.901520 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.902280 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.902849 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.903482 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.904344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gjvgj"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.907709 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.910018 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4z96l"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.911452 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.911951 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.913506 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.914179 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4cxfd"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.915362 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.915748 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.916678 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919828 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9437fcf9-6541-495a-a30c-49c409b40155-serving-cert\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919860 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce18ba60-323b-47c5-8230-81e47fb74a72-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919884 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-config\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919901 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce18ba60-323b-47c5-8230-81e47fb74a72-config\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919917 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221c0f36-52eb-4c35-b5b1-7c59b72caff5-serving-cert\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919933 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67d42404-ab13-451a-84a1-4aba57e7d8bb-proxy-tls\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919952 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-client-ca\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-etcd-serving-ca\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919981 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/957b0629-df4e-473b-9c09-eb9430312c69-node-pullsecrets\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.919995 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-image-import-ca\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920017 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920056 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920074 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-serving-cert\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920116 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920132 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9437fcf9-6541-495a-a30c-49c409b40155-etcd-client\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920147 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78a37133-e23f-4bb2-a42d-b55800e168ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920163 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920177 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957b0629-df4e-473b-9c09-eb9430312c69-audit-dir\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920212 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-etcd-service-ca\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920229 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxghh\" (UniqueName: \"kubernetes.io/projected/1b358a58-bcec-4a77-9031-64915fa2246d-kube-api-access-sxghh\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920244 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/221c0f36-52eb-4c35-b5b1-7c59b72caff5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920262 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1b358a58-bcec-4a77-9031-64915fa2246d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920278 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkh7s\" (UniqueName: \"kubernetes.io/projected/9af337b4-75ac-4e43-8131-73d96f883e16-kube-api-access-gkh7s\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920293 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/49a1ec68-f811-41d8-a01a-1c692341a7ba-auth-proxy-config\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920308 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f8d7281-c6e3-4614-978e-ae5d4c170a85-serving-cert\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/62cf7efa-32a9-4564-875e-b7b6b619805f-images\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920337 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-etcd-client\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920354 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b358a58-bcec-4a77-9031-64915fa2246d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-policies\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920390 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920404 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-encryption-config\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920418 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9w6m\" (UniqueName: \"kubernetes.io/projected/1f8d7281-c6e3-4614-978e-ae5d4c170a85-kube-api-access-t9w6m\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920434 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-proxy-tls\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920452 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r6jb\" (UniqueName: \"kubernetes.io/projected/62cf7efa-32a9-4564-875e-b7b6b619805f-kube-api-access-5r6jb\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920467 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f8d7281-c6e3-4614-978e-ae5d4c170a85-config\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920486 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/683f9d21-7c05-443f-a8c4-d00d9d493db4-srv-cert\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920503 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-config\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920518 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sdvc\" (UniqueName: \"kubernetes.io/projected/49a1ec68-f811-41d8-a01a-1c692341a7ba-kube-api-access-2sdvc\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920537 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vvst\" (UniqueName: \"kubernetes.io/projected/67d42404-ab13-451a-84a1-4aba57e7d8bb-kube-api-access-2vvst\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920559 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkxhg\" (UniqueName: \"kubernetes.io/projected/4c65edd1-18dd-4b95-81c5-c0a4e5d6013d-kube-api-access-lkxhg\") pod \"cluster-samples-operator-665b6dd947-4wrsp\" (UID: \"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920574 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-etcd-ca\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920590 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b358a58-bcec-4a77-9031-64915fa2246d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920606 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t68nc\" (UniqueName: \"kubernetes.io/projected/5809e939-725d-48cb-a582-e115f63873ad-kube-api-access-t68nc\") pod \"dns-operator-744455d44c-l6frd\" (UID: \"5809e939-725d-48cb-a582-e115f63873ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920637 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqvtm\" (UniqueName: \"kubernetes.io/projected/8f3f05cf-48ef-43aa-b938-d67ffdce204e-kube-api-access-nqvtm\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920691 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-config\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920706 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920743 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-serving-cert\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920760 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-config\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920787 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f3f05cf-48ef-43aa-b938-d67ffdce204e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920815 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a9a944-0512-4919-b44f-6109a6aa4aa6-serving-cert\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920834 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-969r6\" (UniqueName: \"kubernetes.io/projected/83a9a944-0512-4919-b44f-6109a6aa4aa6-kube-api-access-969r6\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920850 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c65edd1-18dd-4b95-81c5-c0a4e5d6013d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4wrsp\" (UID: \"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920864 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49a1ec68-f811-41d8-a01a-1c692341a7ba-config\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920880 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920895 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm2jz\" (UniqueName: \"kubernetes.io/projected/221c0f36-52eb-4c35-b5b1-7c59b72caff5-kube-api-access-cm2jz\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920920 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9af337b4-75ac-4e43-8131-73d96f883e16-serving-cert\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920935 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-client-ca\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920949 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920964 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx6cf\" (UniqueName: \"kubernetes.io/projected/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-kube-api-access-vx6cf\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.920999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce18ba60-323b-47c5-8230-81e47fb74a72-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921089 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/62cf7efa-32a9-4564-875e-b7b6b619805f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921118 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6v7\" (UniqueName: \"kubernetes.io/projected/957b0629-df4e-473b-9c09-eb9430312c69-kube-api-access-qr6v7\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921133 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8stl\" (UniqueName: \"kubernetes.io/projected/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-kube-api-access-b8stl\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921148 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78a37133-e23f-4bb2-a42d-b55800e168ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921168 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-etcd-client\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921188 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921202 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/49a1ec68-f811-41d8-a01a-1c692341a7ba-machine-approver-tls\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921219 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-audit\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921258 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62cf7efa-32a9-4564-875e-b7b6b619805f-config\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921273 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-dir\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-serving-cert\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921303 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scqjm\" (UniqueName: \"kubernetes.io/projected/dfcb1f56-7b71-4cf6-a92d-8f174194e697-kube-api-access-scqjm\") pod \"downloads-7954f5f757-dwm2t\" (UID: \"dfcb1f56-7b71-4cf6-a92d-8f174194e697\") " pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-audit-policies\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921335 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921356 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjngr\" (UniqueName: \"kubernetes.io/projected/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-kube-api-access-zjngr\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921387 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-encryption-config\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921403 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fxdk\" (UniqueName: \"kubernetes.io/projected/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-kube-api-access-9fxdk\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921418 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-config\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921432 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8zg9\" (UniqueName: \"kubernetes.io/projected/9437fcf9-6541-495a-a30c-49c409b40155-kube-api-access-c8zg9\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921455 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921470 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67d42404-ab13-451a-84a1-4aba57e7d8bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921490 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78a37133-e23f-4bb2-a42d-b55800e168ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921507 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhbk8\" (UniqueName: \"kubernetes.io/projected/907568c8-1873-41b3-8a58-3651ee7e7922-kube-api-access-hhbk8\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921523 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f8d7281-c6e3-4614-978e-ae5d4c170a85-trusted-ca\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921544 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5809e939-725d-48cb-a582-e115f63873ad-metrics-tls\") pod \"dns-operator-744455d44c-l6frd\" (UID: \"5809e939-725d-48cb-a582-e115f63873ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921559 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg7mr\" (UniqueName: \"kubernetes.io/projected/683f9d21-7c05-443f-a8c4-d00d9d493db4-kube-api-access-tg7mr\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/67d42404-ab13-451a-84a1-4aba57e7d8bb-images\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921594 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-service-ca-bundle\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921610 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921630 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f3f05cf-48ef-43aa-b938-d67ffdce204e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921645 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/683f9d21-7c05-443f-a8c4-d00d9d493db4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921677 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921694 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921711 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921726 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/907568c8-1873-41b3-8a58-3651ee7e7922-audit-dir\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.921741 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.922586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-trusted-ca-bundle\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.923024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-config\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.924695 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49a1ec68-f811-41d8-a01a-1c692341a7ba-config\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.925305 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.928816 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-etcd-client\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.928826 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-client-ca\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.929226 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c65edd1-18dd-4b95-81c5-c0a4e5d6013d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4wrsp\" (UID: \"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.929347 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-etcd-serving-ca\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.929397 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/957b0629-df4e-473b-9c09-eb9430312c69-node-pullsecrets\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.930058 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-image-import-ca\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.930059 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a9a944-0512-4919-b44f-6109a6aa4aa6-serving-cert\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.930967 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.931491 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.931544 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/957b0629-df4e-473b-9c09-eb9430312c69-audit-dir\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.931910 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/221c0f36-52eb-4c35-b5b1-7c59b72caff5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.932155 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b358a58-bcec-4a77-9031-64915fa2246d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.932256 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5ff4f"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.932705 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-audit-policies\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.933215 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.935450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/957b0629-df4e-473b-9c09-eb9430312c69-audit\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.936014 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62cf7efa-32a9-4564-875e-b7b6b619805f-config\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.936060 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-dir\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.936819 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/49a1ec68-f811-41d8-a01a-1c692341a7ba-auth-proxy-config\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.938705 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/62cf7efa-32a9-4564-875e-b7b6b619805f-images\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.938944 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.939026 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4h92h"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.939571 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.939665 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.939075 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.939896 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-service-ca-bundle\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.941040 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.941322 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/907568c8-1873-41b3-8a58-3651ee7e7922-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.941405 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.941619 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.941699 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-serving-cert\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.941905 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9af337b4-75ac-4e43-8131-73d96f883e16-serving-cert\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.942656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.943190 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/907568c8-1873-41b3-8a58-3651ee7e7922-audit-dir\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.943333 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.943414 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9af337b4-75ac-4e43-8131-73d96f883e16-config\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.943760 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-serving-cert\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.944160 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f8d7281-c6e3-4614-978e-ae5d4c170a85-serving-cert\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.944216 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-config\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.944436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/49a1ec68-f811-41d8-a01a-1c692341a7ba-machine-approver-tls\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.944651 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-policies\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.944979 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1b358a58-bcec-4a77-9031-64915fa2246d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.946349 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s4v4x"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.946350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.946634 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f8d7281-c6e3-4614-978e-ae5d4c170a85-trusted-ca\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.946698 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-client-ca\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.946821 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-encryption-config\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.946950 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/957b0629-df4e-473b-9c09-eb9430312c69-etcd-client\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.947006 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/221c0f36-52eb-4c35-b5b1-7c59b72caff5-serving-cert\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.947275 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f8d7281-c6e3-4614-978e-ae5d4c170a85-config\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.947436 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.947565 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-config\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.948307 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.948808 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-serving-cert\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.949532 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.949552 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.949778 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.949968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.950008 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/907568c8-1873-41b3-8a58-3651ee7e7922-encryption-config\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.950143 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8zmq8"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.951588 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.951858 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/62cf7efa-32a9-4564-875e-b7b6b619805f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.952830 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.954077 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.954999 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.955935 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-t6jc6"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.956593 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.957575 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hbl62"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.957732 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.958234 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.958989 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sdqlk"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.960006 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.961091 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dwm2t"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.962222 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.966191 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.966256 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pqnhz"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.966269 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l6frd"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.971129 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.971216 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.971274 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.975389 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.975598 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hbl62"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.975668 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jf9xm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.978238 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.978303 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gr62g"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.978316 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.982123 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t9c2p"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.982200 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4cxfd"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.995061 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.995362 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-brblh"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.996656 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-brblh"] Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.996800 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-brblh" Oct 04 10:52:43 crc kubenswrapper[4758]: I1004 10:52:43.997721 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.012000 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023359 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78a37133-e23f-4bb2-a42d-b55800e168ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023393 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce18ba60-323b-47c5-8230-81e47fb74a72-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023453 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fxdk\" (UniqueName: \"kubernetes.io/projected/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-kube-api-access-9fxdk\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023471 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-config\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023487 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8zg9\" (UniqueName: \"kubernetes.io/projected/9437fcf9-6541-495a-a30c-49c409b40155-kube-api-access-c8zg9\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023512 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5809e939-725d-48cb-a582-e115f63873ad-metrics-tls\") pod \"dns-operator-744455d44c-l6frd\" (UID: \"5809e939-725d-48cb-a582-e115f63873ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023531 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg7mr\" (UniqueName: \"kubernetes.io/projected/683f9d21-7c05-443f-a8c4-d00d9d493db4-kube-api-access-tg7mr\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/67d42404-ab13-451a-84a1-4aba57e7d8bb-images\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023563 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67d42404-ab13-451a-84a1-4aba57e7d8bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023578 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78a37133-e23f-4bb2-a42d-b55800e168ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023601 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f3f05cf-48ef-43aa-b938-d67ffdce204e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023616 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/683f9d21-7c05-443f-a8c4-d00d9d493db4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023632 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9437fcf9-6541-495a-a30c-49c409b40155-serving-cert\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023647 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce18ba60-323b-47c5-8230-81e47fb74a72-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023664 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce18ba60-323b-47c5-8230-81e47fb74a72-config\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023680 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67d42404-ab13-451a-84a1-4aba57e7d8bb-proxy-tls\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023712 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-etcd-service-ca\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023730 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9437fcf9-6541-495a-a30c-49c409b40155-etcd-client\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023744 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78a37133-e23f-4bb2-a42d-b55800e168ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-proxy-tls\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023805 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/683f9d21-7c05-443f-a8c4-d00d9d493db4-srv-cert\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023820 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vvst\" (UniqueName: \"kubernetes.io/projected/67d42404-ab13-451a-84a1-4aba57e7d8bb-kube-api-access-2vvst\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023845 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-etcd-ca\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023871 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t68nc\" (UniqueName: \"kubernetes.io/projected/5809e939-725d-48cb-a582-e115f63873ad-kube-api-access-t68nc\") pod \"dns-operator-744455d44c-l6frd\" (UID: \"5809e939-725d-48cb-a582-e115f63873ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqvtm\" (UniqueName: \"kubernetes.io/projected/8f3f05cf-48ef-43aa-b938-d67ffdce204e-kube-api-access-nqvtm\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.023931 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f3f05cf-48ef-43aa-b938-d67ffdce204e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.024225 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78a37133-e23f-4bb2-a42d-b55800e168ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.024886 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.024896 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-config\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.025066 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f3f05cf-48ef-43aa-b938-d67ffdce204e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.025282 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce18ba60-323b-47c5-8230-81e47fb74a72-config\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.025753 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-etcd-ca\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.026281 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9437fcf9-6541-495a-a30c-49c409b40155-etcd-service-ca\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.026533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/67d42404-ab13-451a-84a1-4aba57e7d8bb-images\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.026799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/67d42404-ab13-451a-84a1-4aba57e7d8bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.028396 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-proxy-tls\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.028881 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f3f05cf-48ef-43aa-b938-d67ffdce204e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.029292 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9437fcf9-6541-495a-a30c-49c409b40155-etcd-client\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.029621 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce18ba60-323b-47c5-8230-81e47fb74a72-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.032857 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.034464 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9437fcf9-6541-495a-a30c-49c409b40155-serving-cert\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.051794 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.068551 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78a37133-e23f-4bb2-a42d-b55800e168ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.072168 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.092265 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.098730 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/67d42404-ab13-451a-84a1-4aba57e7d8bb-proxy-tls\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.111562 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.132203 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.138941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/683f9d21-7c05-443f-a8c4-d00d9d493db4-srv-cert\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.152965 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.159184 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/683f9d21-7c05-443f-a8c4-d00d9d493db4-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.173517 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.193604 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.211777 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.231589 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.251510 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.274237 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.292782 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.313336 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.333841 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.352614 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.372490 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.392498 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.413178 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.419757 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5809e939-725d-48cb-a582-e115f63873ad-metrics-tls\") pod \"dns-operator-744455d44c-l6frd\" (UID: \"5809e939-725d-48cb-a582-e115f63873ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.433204 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.492891 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.514306 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.532170 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.553497 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.573687 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.593685 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.612826 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.633268 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.653134 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.672490 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.692623 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.712930 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.741061 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.752428 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.772447 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.792509 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.812497 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.837087 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.850922 4758 request.go:700] Waited for 1.001195561s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.852207 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.872386 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.892537 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.912953 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.934681 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.952619 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.972246 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 10:52:44 crc kubenswrapper[4758]: I1004 10:52:44.993010 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.013064 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.033054 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.052680 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.086338 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.092414 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.113703 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.132280 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.152761 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.172871 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.193405 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.213305 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.233089 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.252239 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.273384 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.293444 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.312611 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.324924 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.325322 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.325475 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.325827 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.332959 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.352745 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.372151 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.396796 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.412914 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.432495 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.458776 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.472840 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.493175 4758 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.536019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scqjm\" (UniqueName: \"kubernetes.io/projected/dfcb1f56-7b71-4cf6-a92d-8f174194e697-kube-api-access-scqjm\") pod \"downloads-7954f5f757-dwm2t\" (UID: \"dfcb1f56-7b71-4cf6-a92d-8f174194e697\") " pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.560257 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.565156 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-969r6\" (UniqueName: \"kubernetes.io/projected/83a9a944-0512-4919-b44f-6109a6aa4aa6-kube-api-access-969r6\") pod \"controller-manager-879f6c89f-8zvst\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.583688 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6v7\" (UniqueName: \"kubernetes.io/projected/957b0629-df4e-473b-9c09-eb9430312c69-kube-api-access-qr6v7\") pod \"apiserver-76f77b778f-gjvgj\" (UID: \"957b0629-df4e-473b-9c09-eb9430312c69\") " pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.612314 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm2jz\" (UniqueName: \"kubernetes.io/projected/221c0f36-52eb-4c35-b5b1-7c59b72caff5-kube-api-access-cm2jz\") pod \"openshift-config-operator-7777fb866f-4h92h\" (UID: \"221c0f36-52eb-4c35-b5b1-7c59b72caff5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.624640 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.630624 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8stl\" (UniqueName: \"kubernetes.io/projected/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-kube-api-access-b8stl\") pod \"route-controller-manager-6576b87f9c-l9w7c\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.635242 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxghh\" (UniqueName: \"kubernetes.io/projected/1b358a58-bcec-4a77-9031-64915fa2246d-kube-api-access-sxghh\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.653912 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjngr\" (UniqueName: \"kubernetes.io/projected/4de1a2f4-834e-40ba-865d-b9cae56e8ee0-kube-api-access-zjngr\") pod \"openshift-apiserver-operator-796bbdcf4f-6tqhf\" (UID: \"4de1a2f4-834e-40ba-865d-b9cae56e8ee0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.686855 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkh7s\" (UniqueName: \"kubernetes.io/projected/9af337b4-75ac-4e43-8131-73d96f883e16-kube-api-access-gkh7s\") pod \"authentication-operator-69f744f599-pv2mm\" (UID: \"9af337b4-75ac-4e43-8131-73d96f883e16\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.693631 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.709048 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhbk8\" (UniqueName: \"kubernetes.io/projected/907568c8-1873-41b3-8a58-3651ee7e7922-kube-api-access-hhbk8\") pod \"apiserver-7bbb656c7d-rgz5g\" (UID: \"907568c8-1873-41b3-8a58-3651ee7e7922\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.749202 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkxhg\" (UniqueName: \"kubernetes.io/projected/4c65edd1-18dd-4b95-81c5-c0a4e5d6013d-kube-api-access-lkxhg\") pod \"cluster-samples-operator-665b6dd947-4wrsp\" (UID: \"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.776797 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b358a58-bcec-4a77-9031-64915fa2246d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xnclc\" (UID: \"1b358a58-bcec-4a77-9031-64915fa2246d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.777177 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx6cf\" (UniqueName: \"kubernetes.io/projected/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-kube-api-access-vx6cf\") pod \"oauth-openshift-558db77b4-8zmq8\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.779915 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.787930 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9w6m\" (UniqueName: \"kubernetes.io/projected/1f8d7281-c6e3-4614-978e-ae5d4c170a85-kube-api-access-t9w6m\") pod \"console-operator-58897d9998-sdqlk\" (UID: \"1f8d7281-c6e3-4614-978e-ae5d4c170a85\") " pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.795431 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r6jb\" (UniqueName: \"kubernetes.io/projected/62cf7efa-32a9-4564-875e-b7b6b619805f-kube-api-access-5r6jb\") pod \"machine-api-operator-5694c8668f-4z96l\" (UID: \"62cf7efa-32a9-4564-875e-b7b6b619805f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.797028 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.808536 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.812875 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.813216 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.813913 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sdvc\" (UniqueName: \"kubernetes.io/projected/49a1ec68-f811-41d8-a01a-1c692341a7ba-kube-api-access-2sdvc\") pod \"machine-approver-56656f9798-5vx5s\" (UID: \"49a1ec68-f811-41d8-a01a-1c692341a7ba\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.819572 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.823940 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.826982 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.832324 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.848693 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.852357 4758 request.go:700] Waited for 1.89545351s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-tls&limit=500&resourceVersion=0 Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.856705 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.861132 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.872135 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.875990 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dwm2t"] Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.897809 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.908682 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-gjvgj"] Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.913418 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.937646 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.956257 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.975519 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 10:52:45 crc kubenswrapper[4758]: I1004 10:52:45.995075 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.008563 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.020454 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.037409 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fxdk\" (UniqueName: \"kubernetes.io/projected/bf73a85e-8fab-45b4-ad46-4bf96bd00c1b-kube-api-access-9fxdk\") pod \"machine-config-controller-84d6567774-w48jm\" (UID: \"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.047595 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/78a37133-e23f-4bb2-a42d-b55800e168ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm4jl\" (UID: \"78a37133-e23f-4bb2-a42d-b55800e168ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.072244 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.072614 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg7mr\" (UniqueName: \"kubernetes.io/projected/683f9d21-7c05-443f-a8c4-d00d9d493db4-kube-api-access-tg7mr\") pod \"olm-operator-6b444d44fb-lpsnq\" (UID: \"683f9d21-7c05-443f-a8c4-d00d9d493db4\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.099533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce18ba60-323b-47c5-8230-81e47fb74a72-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9k96q\" (UID: \"ce18ba60-323b-47c5-8230-81e47fb74a72\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.109934 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8zmq8"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.115623 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vvst\" (UniqueName: \"kubernetes.io/projected/67d42404-ab13-451a-84a1-4aba57e7d8bb-kube-api-access-2vvst\") pod \"machine-config-operator-74547568cd-n6j98\" (UID: \"67d42404-ab13-451a-84a1-4aba57e7d8bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.118555 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" event={"ID":"49a1ec68-f811-41d8-a01a-1c692341a7ba","Type":"ContainerStarted","Data":"556ed6fbad0ed2834c3fb768a8a747f02105267adfcbf5622191639a02c0cb27"} Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.121014 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dwm2t" event={"ID":"dfcb1f56-7b71-4cf6-a92d-8f174194e697","Type":"ContainerStarted","Data":"140fda099a7596bd5cdc89bed3fba31d0a239a773550a069c86101ae1428f12f"} Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.131215 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" event={"ID":"957b0629-df4e-473b-9c09-eb9430312c69","Type":"ContainerStarted","Data":"0fcaf145277d5328ab0ce74f1d70b66136d99f0e0190dd6f26b10014220c6c80"} Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.137795 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8zg9\" (UniqueName: \"kubernetes.io/projected/9437fcf9-6541-495a-a30c-49c409b40155-kube-api-access-c8zg9\") pod \"etcd-operator-b45778765-5ff4f\" (UID: \"9437fcf9-6541-495a-a30c-49c409b40155\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.165634 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.166410 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t68nc\" (UniqueName: \"kubernetes.io/projected/5809e939-725d-48cb-a582-e115f63873ad-kube-api-access-t68nc\") pod \"dns-operator-744455d44c-l6frd\" (UID: \"5809e939-725d-48cb-a582-e115f63873ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.169219 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.177058 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqvtm\" (UniqueName: \"kubernetes.io/projected/8f3f05cf-48ef-43aa-b938-d67ffdce204e-kube-api-access-nqvtm\") pod \"openshift-controller-manager-operator-756b6f6bc6-s7wx2\" (UID: \"8f3f05cf-48ef-43aa-b938-d67ffdce204e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.187826 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.195650 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.204177 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.204365 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4h92h"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.205076 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.218262 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.218538 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.232658 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.241258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.242753 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pv2mm"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.253918 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 10:52:46 crc kubenswrapper[4758]: W1004 10:52:46.274331 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode43ef621_3b5e_4091_ba7b_ab55ea8d5b4d.slice/crio-487ee9c22b47491629b4e755e88a0fa1fb78199296ff64a7ffbaac9c2891b528 WatchSource:0}: Error finding container 487ee9c22b47491629b4e755e88a0fa1fb78199296ff64a7ffbaac9c2891b528: Status 404 returned error can't find the container with id 487ee9c22b47491629b4e755e88a0fa1fb78199296ff64a7ffbaac9c2891b528 Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.275752 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnzx7\" (UniqueName: \"kubernetes.io/projected/818762b4-81b5-42fe-a8de-5908933306ad-kube-api-access-qnzx7\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283336 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-certificates\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283398 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-tls\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283415 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003701dd-2cdf-4ac6-8835-9017cc03039c-config\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283453 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/818762b4-81b5-42fe-a8de-5908933306ad-service-ca-bundle\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283473 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8km7\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-kube-api-access-c8km7\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-bound-sa-token\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283548 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-default-certificate\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283568 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-stats-auth\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283602 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/003701dd-2cdf-4ac6-8835-9017cc03039c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283626 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-metrics-certs\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283642 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283658 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/003701dd-2cdf-4ac6-8835-9017cc03039c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283704 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283723 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-trusted-ca\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.283740 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.284396 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:46.784381526 +0000 UTC m=+144.077032515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.296495 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.315323 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386558 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386710 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-serving-cert\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386732 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-oauth-config\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386758 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386777 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mbtj\" (UniqueName: \"kubernetes.io/projected/583b0f46-ae37-4ba6-9d33-81fd3b22f633-kube-api-access-5mbtj\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386792 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gffnc\" (UniqueName: \"kubernetes.io/projected/9c1109b9-7685-4315-a3b2-3f3f0d31f861-kube-api-access-gffnc\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386809 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfl4q\" (UniqueName: \"kubernetes.io/projected/a7c28209-920e-49c3-9b2b-334dc06df888-kube-api-access-qfl4q\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386857 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7c28209-920e-49c3-9b2b-334dc06df888-config-volume\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-config\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386927 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-signing-cabundle\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386967 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.386997 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387036 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8klt\" (UniqueName: \"kubernetes.io/projected/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-kube-api-access-s8klt\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387076 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387121 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003701dd-2cdf-4ac6-8835-9017cc03039c-config\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387136 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/73e64143-25b7-4479-b724-8314e9f46069-srv-cert\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387153 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/818762b4-81b5-42fe-a8de-5908933306ad-service-ca-bundle\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387170 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd2nl\" (UniqueName: \"kubernetes.io/projected/37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22-kube-api-access-zd2nl\") pod \"package-server-manager-789f6589d5-snbth\" (UID: \"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387188 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7c28209-920e-49c3-9b2b-334dc06df888-secret-volume\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387202 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4c724dc-0661-4145-b264-eb28df91be33-apiservice-cert\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387268 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8km7\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-kube-api-access-c8km7\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387283 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-service-ca\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387298 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-trusted-ca-bundle\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387337 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-bound-sa-token\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387352 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/73e64143-25b7-4479-b724-8314e9f46069-profile-collector-cert\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387394 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-default-certificate\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387410 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrwh\" (UniqueName: \"kubernetes.io/projected/e6ba4505-f064-4b1d-a545-25e3179485eb-kube-api-access-xmrwh\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mhw4\" (UID: \"e6ba4505-f064-4b1d-a545-25e3179485eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387425 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9c1109b9-7685-4315-a3b2-3f3f0d31f861-certs\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387460 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/003701dd-2cdf-4ac6-8835-9017cc03039c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387476 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1294a413-b496-4780-8ef1-ddd980be9c5c-cert\") pod \"ingress-canary-hbl62\" (UID: \"1294a413-b496-4780-8ef1-ddd980be9c5c\") " pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387491 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-csi-data-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387519 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/003701dd-2cdf-4ac6-8835-9017cc03039c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387534 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-registration-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387564 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-trusted-ca\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387579 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-mountpoint-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387593 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387616 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-socket-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387640 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-config\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387673 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7mhc\" (UniqueName: \"kubernetes.io/projected/aed3108c-fee1-4f60-b31f-ba579cbb34b6-kube-api-access-n7mhc\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387738 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz4md\" (UniqueName: \"kubernetes.io/projected/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-kube-api-access-gz4md\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387752 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-oauth-serving-cert\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387766 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-trusted-ca\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387790 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-snbth\" (UID: \"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387815 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx4wb\" (UniqueName: \"kubernetes.io/projected/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-kube-api-access-cx4wb\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387831 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-certificates\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387848 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnzx7\" (UniqueName: \"kubernetes.io/projected/818762b4-81b5-42fe-a8de-5908933306ad-kube-api-access-qnzx7\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387864 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-signing-key\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387880 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-config-volume\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387904 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww7cs\" (UniqueName: \"kubernetes.io/projected/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-kube-api-access-ww7cs\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387920 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-metrics-tls\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387935 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt728\" (UniqueName: \"kubernetes.io/projected/73e64143-25b7-4479-b724-8314e9f46069-kube-api-access-xt728\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-tls\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387966 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krkpc\" (UniqueName: \"kubernetes.io/projected/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-kube-api-access-krkpc\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.387997 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsdc5\" (UniqueName: \"kubernetes.io/projected/23d17080-6640-4c42-8c07-90b6fe9fa69f-kube-api-access-gsdc5\") pod \"migrator-59844c95c7-j2rqz\" (UID: \"23d17080-6640-4c42-8c07-90b6fe9fa69f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388012 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9c1109b9-7685-4315-a3b2-3f3f0d31f861-node-bootstrap-token\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388080 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/050b26f7-6ca2-4222-b2e0-fa32f256f234-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t9c2p\" (UID: \"050b26f7-6ca2-4222-b2e0-fa32f256f234\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388145 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bbwx\" (UniqueName: \"kubernetes.io/projected/050b26f7-6ca2-4222-b2e0-fa32f256f234-kube-api-access-8bbwx\") pod \"multus-admission-controller-857f4d67dd-t9c2p\" (UID: \"050b26f7-6ca2-4222-b2e0-fa32f256f234\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388173 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-plugins-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f4c724dc-0661-4145-b264-eb28df91be33-tmpfs\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388202 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4c724dc-0661-4145-b264-eb28df91be33-webhook-cert\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388239 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-stats-auth\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388256 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6ba4505-f064-4b1d-a545-25e3179485eb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mhw4\" (UID: \"e6ba4505-f064-4b1d-a545-25e3179485eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388273 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-serving-cert\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388289 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-metrics-certs\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388339 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4str4\" (UniqueName: \"kubernetes.io/projected/ac7d66c4-cd59-4646-8970-5da893df6a56-kube-api-access-4str4\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388416 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhd2s\" (UniqueName: \"kubernetes.io/projected/1294a413-b496-4780-8ef1-ddd980be9c5c-kube-api-access-hhd2s\") pod \"ingress-canary-hbl62\" (UID: \"1294a413-b496-4780-8ef1-ddd980be9c5c\") " pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388431 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-metrics-tls\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.388446 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb9r2\" (UniqueName: \"kubernetes.io/projected/f4c724dc-0661-4145-b264-eb28df91be33-kube-api-access-pb9r2\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.388582 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:46.888559537 +0000 UTC m=+144.181210426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.390000 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.395653 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-default-certificate\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.396795 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-trusted-ca\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.406550 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/818762b4-81b5-42fe-a8de-5908933306ad-service-ca-bundle\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.407617 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/003701dd-2cdf-4ac6-8835-9017cc03039c-config\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.410050 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-certificates\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.410262 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-stats-auth\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.413252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-tls\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.414475 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.422002 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/003701dd-2cdf-4ac6-8835-9017cc03039c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.422146 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/818762b4-81b5-42fe-a8de-5908933306ad-metrics-certs\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.430980 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/003701dd-2cdf-4ac6-8835-9017cc03039c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4vc25\" (UID: \"003701dd-2cdf-4ac6-8835-9017cc03039c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.452223 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnzx7\" (UniqueName: \"kubernetes.io/projected/818762b4-81b5-42fe-a8de-5908933306ad-kube-api-access-qnzx7\") pod \"router-default-5444994796-hcsr7\" (UID: \"818762b4-81b5-42fe-a8de-5908933306ad\") " pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.471756 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8km7\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-kube-api-access-c8km7\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.489919 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-service-ca\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.489955 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-trusted-ca-bundle\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490003 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/73e64143-25b7-4479-b724-8314e9f46069-profile-collector-cert\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490023 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrwh\" (UniqueName: \"kubernetes.io/projected/e6ba4505-f064-4b1d-a545-25e3179485eb-kube-api-access-xmrwh\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mhw4\" (UID: \"e6ba4505-f064-4b1d-a545-25e3179485eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490039 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9c1109b9-7685-4315-a3b2-3f3f0d31f861-certs\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490070 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1294a413-b496-4780-8ef1-ddd980be9c5c-cert\") pod \"ingress-canary-hbl62\" (UID: \"1294a413-b496-4780-8ef1-ddd980be9c5c\") " pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490089 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-csi-data-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490124 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-registration-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490149 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-mountpoint-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490163 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490206 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-socket-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-config\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490238 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7mhc\" (UniqueName: \"kubernetes.io/projected/aed3108c-fee1-4f60-b31f-ba579cbb34b6-kube-api-access-n7mhc\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490279 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz4md\" (UniqueName: \"kubernetes.io/projected/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-kube-api-access-gz4md\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490294 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-oauth-serving-cert\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490311 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-trusted-ca\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490361 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-snbth\" (UID: \"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490379 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx4wb\" (UniqueName: \"kubernetes.io/projected/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-kube-api-access-cx4wb\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490394 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-signing-key\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490409 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-config-volume\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490440 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww7cs\" (UniqueName: \"kubernetes.io/projected/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-kube-api-access-ww7cs\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490454 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-metrics-tls\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490469 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt728\" (UniqueName: \"kubernetes.io/projected/73e64143-25b7-4479-b724-8314e9f46069-kube-api-access-xt728\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490487 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krkpc\" (UniqueName: \"kubernetes.io/projected/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-kube-api-access-krkpc\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490524 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsdc5\" (UniqueName: \"kubernetes.io/projected/23d17080-6640-4c42-8c07-90b6fe9fa69f-kube-api-access-gsdc5\") pod \"migrator-59844c95c7-j2rqz\" (UID: \"23d17080-6640-4c42-8c07-90b6fe9fa69f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490542 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9c1109b9-7685-4315-a3b2-3f3f0d31f861-node-bootstrap-token\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490560 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/050b26f7-6ca2-4222-b2e0-fa32f256f234-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t9c2p\" (UID: \"050b26f7-6ca2-4222-b2e0-fa32f256f234\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bbwx\" (UniqueName: \"kubernetes.io/projected/050b26f7-6ca2-4222-b2e0-fa32f256f234-kube-api-access-8bbwx\") pod \"multus-admission-controller-857f4d67dd-t9c2p\" (UID: \"050b26f7-6ca2-4222-b2e0-fa32f256f234\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490607 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-plugins-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490622 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f4c724dc-0661-4145-b264-eb28df91be33-tmpfs\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490635 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4c724dc-0661-4145-b264-eb28df91be33-webhook-cert\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490673 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6ba4505-f064-4b1d-a545-25e3179485eb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mhw4\" (UID: \"e6ba4505-f064-4b1d-a545-25e3179485eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490690 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-serving-cert\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490712 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4str4\" (UniqueName: \"kubernetes.io/projected/ac7d66c4-cd59-4646-8970-5da893df6a56-kube-api-access-4str4\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490726 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490765 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhd2s\" (UniqueName: \"kubernetes.io/projected/1294a413-b496-4780-8ef1-ddd980be9c5c-kube-api-access-hhd2s\") pod \"ingress-canary-hbl62\" (UID: \"1294a413-b496-4780-8ef1-ddd980be9c5c\") " pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490795 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-metrics-tls\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490839 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb9r2\" (UniqueName: \"kubernetes.io/projected/f4c724dc-0661-4145-b264-eb28df91be33-kube-api-access-pb9r2\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-serving-cert\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490868 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-oauth-config\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490885 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mbtj\" (UniqueName: \"kubernetes.io/projected/583b0f46-ae37-4ba6-9d33-81fd3b22f633-kube-api-access-5mbtj\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490919 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gffnc\" (UniqueName: \"kubernetes.io/projected/9c1109b9-7685-4315-a3b2-3f3f0d31f861-kube-api-access-gffnc\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490935 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfl4q\" (UniqueName: \"kubernetes.io/projected/a7c28209-920e-49c3-9b2b-334dc06df888-kube-api-access-qfl4q\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490952 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7c28209-920e-49c3-9b2b-334dc06df888-config-volume\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-config\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.490999 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-signing-cabundle\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491016 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491064 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8klt\" (UniqueName: \"kubernetes.io/projected/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-kube-api-access-s8klt\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491084 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491118 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/73e64143-25b7-4479-b724-8314e9f46069-srv-cert\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491135 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd2nl\" (UniqueName: \"kubernetes.io/projected/37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22-kube-api-access-zd2nl\") pod \"package-server-manager-789f6589d5-snbth\" (UID: \"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491151 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7c28209-920e-49c3-9b2b-334dc06df888-secret-volume\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.491165 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4c724dc-0661-4145-b264-eb28df91be33-apiservice-cert\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.492472 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.492706 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:46.992695926 +0000 UTC m=+144.285346815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.492793 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-oauth-serving-cert\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.493335 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-trusted-ca\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.497200 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-service-ca\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.497983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-trusted-ca-bundle\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.498538 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-config\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.498851 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-signing-cabundle\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.499036 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7c28209-920e-49c3-9b2b-334dc06df888-config-volume\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.499786 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.501554 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-config-volume\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.506056 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4z96l"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.506954 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f4c724dc-0661-4145-b264-eb28df91be33-tmpfs\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.507280 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-plugins-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.507693 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7c28209-920e-49c3-9b2b-334dc06df888-secret-volume\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.509893 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/050b26f7-6ca2-4222-b2e0-fa32f256f234-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t9c2p\" (UID: \"050b26f7-6ca2-4222-b2e0-fa32f256f234\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.510490 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-metrics-tls\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.510928 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-snbth\" (UID: \"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.510940 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9c1109b9-7685-4315-a3b2-3f3f0d31f861-node-bootstrap-token\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.511398 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-registration-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.511482 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-bound-sa-token\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.511508 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/73e64143-25b7-4479-b724-8314e9f46069-srv-cert\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.511548 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-socket-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.511586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-mountpoint-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.511850 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-metrics-tls\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.512557 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4c724dc-0661-4145-b264-eb28df91be33-webhook-cert\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.513429 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-config\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.513533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ac7d66c4-cd59-4646-8970-5da893df6a56-csi-data-dir\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.516226 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-signing-key\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.517680 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4c724dc-0661-4145-b264-eb28df91be33-apiservice-cert\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.518127 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-serving-cert\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.522887 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/73e64143-25b7-4479-b724-8314e9f46069-profile-collector-cert\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.523523 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.523710 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-serving-cert\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.525044 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1294a413-b496-4780-8ef1-ddd980be9c5c-cert\") pod \"ingress-canary-hbl62\" (UID: \"1294a413-b496-4780-8ef1-ddd980be9c5c\") " pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.526664 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9c1109b9-7685-4315-a3b2-3f3f0d31f861-certs\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.531168 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.535065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.536734 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6ba4505-f064-4b1d-a545-25e3179485eb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mhw4\" (UID: \"e6ba4505-f064-4b1d-a545-25e3179485eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.550156 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-oauth-config\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.550940 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8zvst"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.553547 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz4md\" (UniqueName: \"kubernetes.io/projected/11fe31d1-ad44-4482-ba21-7c0c8a4d8a03-kube-api-access-gz4md\") pod \"service-ca-9c57cc56f-jf9xm\" (UID: \"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03\") " pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.560586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4str4\" (UniqueName: \"kubernetes.io/projected/ac7d66c4-cd59-4646-8970-5da893df6a56-kube-api-access-4str4\") pod \"csi-hostpathplugin-4cxfd\" (UID: \"ac7d66c4-cd59-4646-8970-5da893df6a56\") " pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.566131 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfl4q\" (UniqueName: \"kubernetes.io/projected/a7c28209-920e-49c3-9b2b-334dc06df888-kube-api-access-qfl4q\") pod \"collect-profiles-29326245-p5jpm\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.574964 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.587713 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.594414 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.595970 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.095953194 +0000 UTC m=+144.388604083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.601714 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krkpc\" (UniqueName: \"kubernetes.io/projected/1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b-kube-api-access-krkpc\") pod \"kube-storage-version-migrator-operator-b67b599dd-qq482\" (UID: \"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.606778 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.608734 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.619059 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsdc5\" (UniqueName: \"kubernetes.io/projected/23d17080-6640-4c42-8c07-90b6fe9fa69f-kube-api-access-gsdc5\") pod \"migrator-59844c95c7-j2rqz\" (UID: \"23d17080-6640-4c42-8c07-90b6fe9fa69f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.624253 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.639591 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhd2s\" (UniqueName: \"kubernetes.io/projected/1294a413-b496-4780-8ef1-ddd980be9c5c-kube-api-access-hhd2s\") pod \"ingress-canary-hbl62\" (UID: \"1294a413-b496-4780-8ef1-ddd980be9c5c\") " pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.652189 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb9r2\" (UniqueName: \"kubernetes.io/projected/f4c724dc-0661-4145-b264-eb28df91be33-kube-api-access-pb9r2\") pod \"packageserver-d55dfcdfc-tfqb6\" (UID: \"f4c724dc-0661-4145-b264-eb28df91be33\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.673495 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx4wb\" (UniqueName: \"kubernetes.io/projected/0b57bdb3-1a08-4a62-b4c2-ee25c470f84d-kube-api-access-cx4wb\") pod \"service-ca-operator-777779d784-gpsj2\" (UID: \"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.695053 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.696620 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.696887 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.196876482 +0000 UTC m=+144.489527371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.706810 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww7cs\" (UniqueName: \"kubernetes.io/projected/fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5-kube-api-access-ww7cs\") pod \"dns-default-brblh\" (UID: \"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5\") " pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.714125 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hbl62" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.716932 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-brblh" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.734751 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt728\" (UniqueName: \"kubernetes.io/projected/73e64143-25b7-4479-b724-8314e9f46069-kube-api-access-xt728\") pod \"catalog-operator-68c6474976-vdnpz\" (UID: \"73e64143-25b7-4479-b724-8314e9f46069\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.765745 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd2nl\" (UniqueName: \"kubernetes.io/projected/37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22-kube-api-access-zd2nl\") pod \"package-server-manager-789f6589d5-snbth\" (UID: \"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.768595 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bbwx\" (UniqueName: \"kubernetes.io/projected/050b26f7-6ca2-4222-b2e0-fa32f256f234-kube-api-access-8bbwx\") pod \"multus-admission-controller-857f4d67dd-t9c2p\" (UID: \"050b26f7-6ca2-4222-b2e0-fa32f256f234\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.779016 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.800333 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.800795 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.300778116 +0000 UTC m=+144.593429005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.804429 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sdqlk"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.809956 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp"] Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.829344 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gffnc\" (UniqueName: \"kubernetes.io/projected/9c1109b9-7685-4315-a3b2-3f3f0d31f861-kube-api-access-gffnc\") pod \"machine-config-server-t6jc6\" (UID: \"9c1109b9-7685-4315-a3b2-3f3f0d31f861\") " pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.833985 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mbtj\" (UniqueName: \"kubernetes.io/projected/583b0f46-ae37-4ba6-9d33-81fd3b22f633-kube-api-access-5mbtj\") pod \"marketplace-operator-79b997595-s4v4x\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.847980 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.863410 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.863598 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrwh\" (UniqueName: \"kubernetes.io/projected/e6ba4505-f064-4b1d-a545-25e3179485eb-kube-api-access-xmrwh\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mhw4\" (UID: \"e6ba4505-f064-4b1d-a545-25e3179485eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.875331 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8klt\" (UniqueName: \"kubernetes.io/projected/c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1-kube-api-access-s8klt\") pod \"ingress-operator-5b745b69d9-2vd2f\" (UID: \"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.879719 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.889356 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.890418 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7mhc\" (UniqueName: \"kubernetes.io/projected/aed3108c-fee1-4f60-b31f-ba579cbb34b6-kube-api-access-n7mhc\") pod \"console-f9d7485db-pqnhz\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.895925 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.904864 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:46 crc kubenswrapper[4758]: E1004 10:52:46.905165 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.405151891 +0000 UTC m=+144.697802780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.915162 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.945433 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.952173 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.967477 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.967888 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" Oct 04 10:52:46 crc kubenswrapper[4758]: I1004 10:52:46.971713 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.011904 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.012281 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.512266196 +0000 UTC m=+144.804917085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.012357 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-t6jc6" Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.024982 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.054052 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5ff4f"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.065305 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.113279 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.113616 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.613605155 +0000 UTC m=+144.906256044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.166634 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.171517 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.205485 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" event={"ID":"83a9a944-0512-4919-b44f-6109a6aa4aa6","Type":"ContainerStarted","Data":"76316df3115629407fa3af892b96f6de53781f5636cc9a81aec9d615daf7c3d3"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.219988 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.220254 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.720241237 +0000 UTC m=+145.012892126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.274263 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dwm2t" event={"ID":"dfcb1f56-7b71-4cf6-a92d-8f174194e697","Type":"ContainerStarted","Data":"90b097e1e0437a029a25c720b271d4104f28abda2db22d6c910861617b2f4135"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.275045 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.277997 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-dwm2t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.278051 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dwm2t" podUID="dfcb1f56-7b71-4cf6-a92d-8f174194e697" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.279281 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l6frd"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.280537 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" event={"ID":"1b358a58-bcec-4a77-9031-64915fa2246d","Type":"ContainerStarted","Data":"758e63d48df33ce32f6669e4daf1a251593f9d9fb25bc80d5dcda87c415a5ed6"} Oct 04 10:52:47 crc kubenswrapper[4758]: W1004 10:52:47.280738 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod818762b4_81b5_42fe_a8de_5908933306ad.slice/crio-a0bfa354a46fac9702438b7494d76cd9d46b3c75862e0a601a0d1ed1506672ce WatchSource:0}: Error finding container a0bfa354a46fac9702438b7494d76cd9d46b3c75862e0a601a0d1ed1506672ce: Status 404 returned error can't find the container with id a0bfa354a46fac9702438b7494d76cd9d46b3c75862e0a601a0d1ed1506672ce Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.284507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" event={"ID":"4de1a2f4-834e-40ba-865d-b9cae56e8ee0","Type":"ContainerStarted","Data":"62e44e09f928f9bc3ae08f346bf6679879fa15f926b0c19f00fa74f501d36c21"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.288997 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.290254 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.305143 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" event={"ID":"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9","Type":"ContainerStarted","Data":"ce042da7d217ddadaa3c048758d056f6d643483a8c5912f365a593d66b5d25f6"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.307738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" event={"ID":"49a1ec68-f811-41d8-a01a-1c692341a7ba","Type":"ContainerStarted","Data":"a2f92bf71e7a26a09eedc41fbfa1dacb8118fce76786b2e8eed4df74bf3653af"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.310745 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.322516 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.322950 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.82293092 +0000 UTC m=+145.115581889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.325242 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" event={"ID":"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d","Type":"ContainerStarted","Data":"487ee9c22b47491629b4e755e88a0fa1fb78199296ff64a7ffbaac9c2891b528"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.330745 4758 generic.go:334] "Generic (PLEG): container finished" podID="957b0629-df4e-473b-9c09-eb9430312c69" containerID="2232941321a305e7eb1facd6eb6dffb53df46ff88e9836200e4093facad86252" exitCode=0 Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.356596 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" event={"ID":"957b0629-df4e-473b-9c09-eb9430312c69","Type":"ContainerDied","Data":"2232941321a305e7eb1facd6eb6dffb53df46ff88e9836200e4093facad86252"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.356635 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" event={"ID":"9af337b4-75ac-4e43-8131-73d96f883e16","Type":"ContainerStarted","Data":"45ee63133bbd05d54c7e000416d5bc6f6c6b54da92a23c709b34f1dc9787102a"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.356645 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" event={"ID":"9af337b4-75ac-4e43-8131-73d96f883e16","Type":"ContainerStarted","Data":"2e1d933f2f4521c0e7e3d943d2884a19026c3af8ffa613cbb57062c71138912f"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.367082 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" event={"ID":"907568c8-1873-41b3-8a58-3651ee7e7922","Type":"ContainerStarted","Data":"356d8ee0919417dc0b493dd6bb7a791960f85f79df9c6ae4fc5e984db0e9a4df"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.369453 4758 generic.go:334] "Generic (PLEG): container finished" podID="221c0f36-52eb-4c35-b5b1-7c59b72caff5" containerID="bf8ea3bd040084f262d735b7ef3e16ac627f542c79e2e2f37f7f461247e60844" exitCode=0 Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.369496 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" event={"ID":"221c0f36-52eb-4c35-b5b1-7c59b72caff5","Type":"ContainerDied","Data":"bf8ea3bd040084f262d735b7ef3e16ac627f542c79e2e2f37f7f461247e60844"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.369513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" event={"ID":"221c0f36-52eb-4c35-b5b1-7c59b72caff5","Type":"ContainerStarted","Data":"fa7218903293cedce97d1fa6b91c3d10446fb0a6cb5fbfe14801e9a16b6ef8c6"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.370719 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" event={"ID":"62cf7efa-32a9-4564-875e-b7b6b619805f","Type":"ContainerStarted","Data":"65e47da88895ec860e6b26d160030191c33bb0081679a1c492b6168c09a47403"} Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.376035 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" event={"ID":"1f8d7281-c6e3-4614-978e-ae5d4c170a85","Type":"ContainerStarted","Data":"9a09dc6238fbd98a81bdcd3d78d6c757370838287189099bffe493f081fa0b01"} Oct 04 10:52:47 crc kubenswrapper[4758]: W1004 10:52:47.392085 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5809e939_725d_48cb_a582_e115f63873ad.slice/crio-e5fac1618f8a8c334658d83a6035d1412ed13452ce011351e0e9efc2fb8ab7bc WatchSource:0}: Error finding container e5fac1618f8a8c334658d83a6035d1412ed13452ce011351e0e9efc2fb8ab7bc: Status 404 returned error can't find the container with id e5fac1618f8a8c334658d83a6035d1412ed13452ce011351e0e9efc2fb8ab7bc Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.423757 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.425160 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:47.925145422 +0000 UTC m=+145.217796311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.500745 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.528592 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.563224 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.063207548 +0000 UTC m=+145.355858437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.594144 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jf9xm"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.630144 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.630483 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.130469386 +0000 UTC m=+145.423120275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: W1004 10:52:47.636526 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod003701dd_2cdf_4ac6_8835_9017cc03039c.slice/crio-8a7aa743626003906222865e90e8d3a1f042026983a7d0d8893821eb90c77504 WatchSource:0}: Error finding container 8a7aa743626003906222865e90e8d3a1f042026983a7d0d8893821eb90c77504: Status 404 returned error can't find the container with id 8a7aa743626003906222865e90e8d3a1f042026983a7d0d8893821eb90c77504 Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.641056 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hbl62"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.646291 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-brblh"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.649779 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-4cxfd"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.649827 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.698477 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pv2mm" podStartSLOduration=122.698459333 podStartE2EDuration="2m2.698459333s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:47.688647705 +0000 UTC m=+144.981298594" watchObservedRunningTime="2025-10-04 10:52:47.698459333 +0000 UTC m=+144.991110222" Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.700560 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.732188 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.733231 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.23321444 +0000 UTC m=+145.525865329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.764055 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz"] Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.833087 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.833300 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.333254647 +0000 UTC m=+145.625905526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.833381 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.833739 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.333727899 +0000 UTC m=+145.626378788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.890668 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dwm2t" podStartSLOduration=122.890646026 podStartE2EDuration="2m2.890646026s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:47.882789357 +0000 UTC m=+145.175440246" watchObservedRunningTime="2025-10-04 10:52:47.890646026 +0000 UTC m=+145.183296915" Oct 04 10:52:47 crc kubenswrapper[4758]: W1004 10:52:47.909175 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23d17080_6640_4c42_8c07_90b6fe9fa69f.slice/crio-0a076e87c478570603ecbec47375852eb6de7fd1c8f2569769f944d2be67454f WatchSource:0}: Error finding container 0a076e87c478570603ecbec47375852eb6de7fd1c8f2569769f944d2be67454f: Status 404 returned error can't find the container with id 0a076e87c478570603ecbec47375852eb6de7fd1c8f2569769f944d2be67454f Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.934162 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.934330 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.434307648 +0000 UTC m=+145.726958527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:47 crc kubenswrapper[4758]: I1004 10:52:47.934422 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:47 crc kubenswrapper[4758]: E1004 10:52:47.934742 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.434730529 +0000 UTC m=+145.727381418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.036250 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.036407 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.536384386 +0000 UTC m=+145.829035275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.038571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.039060 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.539045463 +0000 UTC m=+145.831696352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.142555 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.142897 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.642881805 +0000 UTC m=+145.935532694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.153138 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.263369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.264732 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.764700081 +0000 UTC m=+146.057350970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.312325 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.365616 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.366021 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.866006349 +0000 UTC m=+146.158657238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: W1004 10:52:48.355222 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fe83ad0_c0fa_4b95_9fbf_111dc0b6df0b.slice/crio-d05219b331f2d498d9c7b967dcb15fb7ce36c5a8a53ea7991c793769c0febc08 WatchSource:0}: Error finding container d05219b331f2d498d9c7b967dcb15fb7ce36c5a8a53ea7991c793769c0febc08: Status 404 returned error can't find the container with id d05219b331f2d498d9c7b967dcb15fb7ce36c5a8a53ea7991c793769c0febc08 Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.379389 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t9c2p"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.448616 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.468759 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.469047 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:48.969036871 +0000 UTC m=+146.261687760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.474707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" event={"ID":"221c0f36-52eb-4c35-b5b1-7c59b72caff5","Type":"ContainerStarted","Data":"71e45931f5b09358466109416c80a6120ba2c82043eead5844c28acdad41313c"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.475148 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.494663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" event={"ID":"ce18ba60-323b-47c5-8230-81e47fb74a72","Type":"ContainerStarted","Data":"47a934a325b8ce61f01fa8461f4341002cfb7ebf3d17556ef5936e5a77fc8a86"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.499628 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-t6jc6" event={"ID":"9c1109b9-7685-4315-a3b2-3f3f0d31f861","Type":"ContainerStarted","Data":"a4300177a05066e3f8ca8a0f7ba3efac8cdd01b751e6c62e4bfa8fd6d13e31a0"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.522916 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s4v4x"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.522965 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.557663 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.557707 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-pqnhz"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.571118 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f"] Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.571554 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" event={"ID":"683f9d21-7c05-443f-a8c4-d00d9d493db4","Type":"ContainerStarted","Data":"7804bf1f31683855455e48eea93b1042a80ac5fda2380faa430b895a791239e4"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.571600 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" event={"ID":"683f9d21-7c05-443f-a8c4-d00d9d493db4","Type":"ContainerStarted","Data":"3d185eb1e92b41ac90ccdfeca57e7ad043b553a1f1c5289a5eb679a8977ac8a2"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.572075 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.572247 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.573059 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.073043527 +0000 UTC m=+146.365694416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.575526 4758 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-lpsnq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.575593 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" podUID="683f9d21-7c05-443f-a8c4-d00d9d493db4" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.576502 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" podStartSLOduration=123.576492174 podStartE2EDuration="2m3.576492174s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:48.575979231 +0000 UTC m=+145.868630120" watchObservedRunningTime="2025-10-04 10:52:48.576492174 +0000 UTC m=+145.869143053" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.635236 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" event={"ID":"4de1a2f4-834e-40ba-865d-b9cae56e8ee0","Type":"ContainerStarted","Data":"1f84410fab08baa2e253b5d1b7c46a895c4f57637a178577a09384657d39ffe6"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.668069 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" podStartSLOduration=123.668055015 podStartE2EDuration="2m3.668055015s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:48.613924618 +0000 UTC m=+145.906575507" watchObservedRunningTime="2025-10-04 10:52:48.668055015 +0000 UTC m=+145.960705904" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.669761 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" event={"ID":"5809e939-725d-48cb-a582-e115f63873ad","Type":"ContainerStarted","Data":"e5fac1618f8a8c334658d83a6035d1412ed13452ce011351e0e9efc2fb8ab7bc"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.674782 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.675799 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.17578788 +0000 UTC m=+146.468438769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.699687 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" event={"ID":"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b","Type":"ContainerStarted","Data":"2be5cc570d9e4656639ef3505668cce268611de2561e96ddbdb4905b6dcb5c42"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.726955 4758 generic.go:334] "Generic (PLEG): container finished" podID="907568c8-1873-41b3-8a58-3651ee7e7922" containerID="d07179ad200b61ba908203dbfd0fce63e45e87ed6753884e059ec0f787cee098" exitCode=0 Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.727189 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" event={"ID":"907568c8-1873-41b3-8a58-3651ee7e7922","Type":"ContainerDied","Data":"d07179ad200b61ba908203dbfd0fce63e45e87ed6753884e059ec0f787cee098"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.763339 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6tqhf" podStartSLOduration=123.76331966 podStartE2EDuration="2m3.76331966s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:48.669420559 +0000 UTC m=+145.962071448" watchObservedRunningTime="2025-10-04 10:52:48.76331966 +0000 UTC m=+146.055970549" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.785202 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.786253 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.286221909 +0000 UTC m=+146.578872798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.800496 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" event={"ID":"62cf7efa-32a9-4564-875e-b7b6b619805f","Type":"ContainerStarted","Data":"35dae1189bdcff89137a54c693cf5fcaa54d4bb9730aa666864cb7bdb7b130ef"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.825027 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-brblh" event={"ID":"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5","Type":"ContainerStarted","Data":"4235794860a24745d0a63d5fbcde1b60dc3f462b0bacb89a9245ca020d1133f6"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.829273 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" event={"ID":"78a37133-e23f-4bb2-a42d-b55800e168ca","Type":"ContainerStarted","Data":"7a3a1c74858080d62ec27edb7ba3b172aaa0b526b7e53491fbec4c480eea90e6"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.833792 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" event={"ID":"ac7d66c4-cd59-4646-8970-5da893df6a56","Type":"ContainerStarted","Data":"daec54d859b14c794b22442c3f0375053084a7b0e992b241308deeda97e917fa"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.853941 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" event={"ID":"23d17080-6640-4c42-8c07-90b6fe9fa69f","Type":"ContainerStarted","Data":"0a076e87c478570603ecbec47375852eb6de7fd1c8f2569769f944d2be67454f"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.859891 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" event={"ID":"1f8d7281-c6e3-4614-978e-ae5d4c170a85","Type":"ContainerStarted","Data":"e5cd70d9fdf3a94964ca2b2e848abe108680903afa63382f25a93dfd98167a71"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.860866 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.863274 4758 patch_prober.go:28] interesting pod/console-operator-58897d9998-sdqlk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.863305 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" podUID="1f8d7281-c6e3-4614-978e-ae5d4c170a85" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.873933 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hbl62" event={"ID":"1294a413-b496-4780-8ef1-ddd980be9c5c","Type":"ContainerStarted","Data":"7e5c132ec09bd65bcd0bc81603279ffa03445a5160b93e83a0c30022243c1f8e"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.876566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" event={"ID":"83a9a944-0512-4919-b44f-6109a6aa4aa6","Type":"ContainerStarted","Data":"ebc4970a2f2e2b2deaf52766c3adf17759ea58f511635edbae448d2a7b26a386"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.877375 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.885239 4758 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8zvst container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.885291 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" podUID="83a9a944-0512-4919-b44f-6109a6aa4aa6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.886951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.887796 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.387781723 +0000 UTC m=+146.680432722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.892252 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" podStartSLOduration=123.892230885 podStartE2EDuration="2m3.892230885s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:48.885778443 +0000 UTC m=+146.178429332" watchObservedRunningTime="2025-10-04 10:52:48.892230885 +0000 UTC m=+146.184881774" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.894709 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" event={"ID":"003701dd-2cdf-4ac6-8835-9017cc03039c","Type":"ContainerStarted","Data":"8a7aa743626003906222865e90e8d3a1f042026983a7d0d8893821eb90c77504"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.910093 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" event={"ID":"1b358a58-bcec-4a77-9031-64915fa2246d","Type":"ContainerStarted","Data":"ba56dde164050165f798ba861af3c16e2b4fa13d30e073b0a5f7ba76741802e4"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.924361 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" podStartSLOduration=123.924343906 podStartE2EDuration="2m3.924343906s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:48.908614849 +0000 UTC m=+146.201265738" watchObservedRunningTime="2025-10-04 10:52:48.924343906 +0000 UTC m=+146.216994795" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.944595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" event={"ID":"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03","Type":"ContainerStarted","Data":"de533850da146b6f9eae57a3b3cc1894363abaa02c2d5dee6ed8120bbd2a2af2"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.975005 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" event={"ID":"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d","Type":"ContainerStarted","Data":"eb6080f3c55f04d17a08636d307b7a3d36f2e2a2ea4bd74f98d45150fdf45cc8"} Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.975616 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hbl62" podStartSLOduration=5.975593941 podStartE2EDuration="5.975593941s" podCreationTimestamp="2025-10-04 10:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:48.968967763 +0000 UTC m=+146.261618652" watchObservedRunningTime="2025-10-04 10:52:48.975593941 +0000 UTC m=+146.268244830" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.977172 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:48 crc kubenswrapper[4758]: I1004 10:52:48.990485 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:48 crc kubenswrapper[4758]: E1004 10:52:48.992729 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.492708233 +0000 UTC m=+146.785359122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:48.996622 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" event={"ID":"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d","Type":"ContainerStarted","Data":"b7bedda4b0c563ecee6d5778ad3705e4b23b46aa6d4fd339610277bc5d8e2148"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:48.996672 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" event={"ID":"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d","Type":"ContainerStarted","Data":"2cae87c99adda932c50df2540bb18e5eb0ff4dd30a5322f7408ab4be3df5b83c"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.001194 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.002275 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.502243163 +0000 UTC m=+146.794894052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.015535 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" podStartSLOduration=124.015498668 podStartE2EDuration="2m4.015498668s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.014521473 +0000 UTC m=+146.307172352" watchObservedRunningTime="2025-10-04 10:52:49.015498668 +0000 UTC m=+146.308149557" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.021924 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" event={"ID":"8f3f05cf-48ef-43aa-b938-d67ffdce204e","Type":"ContainerStarted","Data":"b4511eed5a7181116eb344fc4a791bf9ed44d9e573d2a290597c4c54f339987f"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.048256 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xnclc" podStartSLOduration=124.048240425 podStartE2EDuration="2m4.048240425s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.045770462 +0000 UTC m=+146.338421351" watchObservedRunningTime="2025-10-04 10:52:49.048240425 +0000 UTC m=+146.340891304" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.048780 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" event={"ID":"67d42404-ab13-451a-84a1-4aba57e7d8bb","Type":"ContainerStarted","Data":"0bd9dfeb6556451374924a52cc99f4a4015410f23211fbd0703fac651b10b65c"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.072818 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" event={"ID":"49a1ec68-f811-41d8-a01a-1c692341a7ba","Type":"ContainerStarted","Data":"cfa0577dd5cf48ba9967f1ee19f96fda88e546630eb3cbb8e7012a6faf83e587"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.102769 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.104215 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.604196048 +0000 UTC m=+146.896846937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.108590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" event={"ID":"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9","Type":"ContainerStarted","Data":"1f52a7d951dc8528a34a49e7af719242b46acfe1f9bbaab97b1cc49646199003"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.109372 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.130745 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" event={"ID":"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d","Type":"ContainerStarted","Data":"b4c1252e5b8923e23291a4df3a11babe8fcc304b3a704275047b0576ba50d1b8"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.147957 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" podStartSLOduration=124.147942022 podStartE2EDuration="2m4.147942022s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.102191777 +0000 UTC m=+146.394842666" watchObservedRunningTime="2025-10-04 10:52:49.147942022 +0000 UTC m=+146.440592911" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.162493 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" event={"ID":"9437fcf9-6541-495a-a30c-49c409b40155","Type":"ContainerStarted","Data":"cea934a8f953c8c770cfb031222c9489e51f10464af230285e059b44f23a81a1"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.182349 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" event={"ID":"a7c28209-920e-49c3-9b2b-334dc06df888","Type":"ContainerStarted","Data":"007a0a0ea970e919e1b2573ca0af6a1fbd7eec539fe95843b58b0f2ca5070c6b"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.206239 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" podStartSLOduration=124.206219144 podStartE2EDuration="2m4.206219144s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.153270077 +0000 UTC m=+146.445920966" watchObservedRunningTime="2025-10-04 10:52:49.206219144 +0000 UTC m=+146.498870033" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.208040 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.208336 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.708323327 +0000 UTC m=+147.000974216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.218741 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hcsr7" event={"ID":"818762b4-81b5-42fe-a8de-5908933306ad","Type":"ContainerStarted","Data":"9ee47c95780f30bdd69fed093b75675499f8970a20875181432eab72cd690cc6"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.218793 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hcsr7" event={"ID":"818762b4-81b5-42fe-a8de-5908933306ad","Type":"ContainerStarted","Data":"a0bfa354a46fac9702438b7494d76cd9d46b3c75862e0a601a0d1ed1506672ce"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.223016 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" event={"ID":"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b","Type":"ContainerStarted","Data":"d05219b331f2d498d9c7b967dcb15fb7ce36c5a8a53ea7991c793769c0febc08"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.244301 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vx5s" podStartSLOduration=124.244280495 podStartE2EDuration="2m4.244280495s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.238761336 +0000 UTC m=+146.531412225" watchObservedRunningTime="2025-10-04 10:52:49.244280495 +0000 UTC m=+146.536931404" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.244798 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" podStartSLOduration=124.244793108 podStartE2EDuration="2m4.244793108s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.20726193 +0000 UTC m=+146.499912819" watchObservedRunningTime="2025-10-04 10:52:49.244793108 +0000 UTC m=+146.537443997" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.262563 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" event={"ID":"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22","Type":"ContainerStarted","Data":"3315ffcf66a000387ba501ca6e993c38f8ec40feb338354d625be04d28152ea2"} Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.263227 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-dwm2t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.263255 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dwm2t" podUID="dfcb1f56-7b71-4cf6-a92d-8f174194e697" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.263625 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" podStartSLOduration=124.263608863 podStartE2EDuration="2m4.263608863s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.262230828 +0000 UTC m=+146.554881717" watchObservedRunningTime="2025-10-04 10:52:49.263608863 +0000 UTC m=+146.556259752" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.303721 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-hcsr7" podStartSLOduration=124.303704776 podStartE2EDuration="2m4.303704776s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:49.303178622 +0000 UTC m=+146.595829511" watchObservedRunningTime="2025-10-04 10:52:49.303704776 +0000 UTC m=+146.596355665" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.311240 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.314061 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.814042087 +0000 UTC m=+147.106692976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.415746 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.417860 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:49.917849248 +0000 UTC m=+147.210500137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.521647 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.521925 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.021910135 +0000 UTC m=+147.314561024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.541506 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.551513 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:49 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:49 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:49 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.551558 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.626836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.627393 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.127371978 +0000 UTC m=+147.420022867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.635064 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.728317 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.728998 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.228980934 +0000 UTC m=+147.521631813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.835541 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.835875 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.335863353 +0000 UTC m=+147.628514242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.920281 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:52:49 crc kubenswrapper[4758]: I1004 10:52:49.944294 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:49 crc kubenswrapper[4758]: E1004 10:52:49.944808 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.444793013 +0000 UTC m=+147.737443902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.052351 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.052644 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.552632936 +0000 UTC m=+147.845283825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.154660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.155020 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.655005201 +0000 UTC m=+147.947656090 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.256951 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.257543 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.75752897 +0000 UTC m=+148.050179859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.320558 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hbl62" event={"ID":"1294a413-b496-4780-8ef1-ddd980be9c5c","Type":"ContainerStarted","Data":"d7ed91fe5a7f41c3f8b03c998dd6eb079967f092e7098c9382828b1c6c20cabb"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.327389 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" event={"ID":"67d42404-ab13-451a-84a1-4aba57e7d8bb","Type":"ContainerStarted","Data":"48dd4163c3d09872b60306e949b62f59a973ca40b436418f0bebb6bd21ef3ae9"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.358473 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.360399 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.860375087 +0000 UTC m=+148.153025986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.362736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" event={"ID":"f4c724dc-0661-4145-b264-eb28df91be33","Type":"ContainerStarted","Data":"cd514fd1523e322a8bd871b49791fdebe3f2e69d2435c4261e95eebcd809e756"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.363076 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" event={"ID":"f4c724dc-0661-4145-b264-eb28df91be33","Type":"ContainerStarted","Data":"3b4d94e701d9d9480c2d750719cb1fb4afb7a02e5635c7e2e636744f6b51ee90"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.364138 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.365937 4758 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tfqb6 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.366009 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" podUID="f4c724dc-0661-4145-b264-eb28df91be33" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.393002 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-s7wx2" event={"ID":"8f3f05cf-48ef-43aa-b938-d67ffdce204e","Type":"ContainerStarted","Data":"b139b57b55b3c1169d32d239ab85b6723f80a591e8faaf146572c7e447ad6294"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.396607 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" event={"ID":"73e64143-25b7-4479-b724-8314e9f46069","Type":"ContainerStarted","Data":"9890c884cc6005672e6efa3ca509634c52ff34e99a004a456aa169a5e9d5fdb6"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.396627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" event={"ID":"73e64143-25b7-4479-b724-8314e9f46069","Type":"ContainerStarted","Data":"bb67dfec264ba323d748f2aaa5f8fc4bc55db0c26bc5df19bacedfe8c6f59b47"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.397194 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.398207 4758 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vdnpz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.398235 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" podUID="73e64143-25b7-4479-b724-8314e9f46069" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.408820 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" event={"ID":"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1","Type":"ContainerStarted","Data":"52fd11248cd88647f78dacca92971a4f4f4c2f30ea57ab58cff726f50db02e7f"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.408855 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" event={"ID":"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1","Type":"ContainerStarted","Data":"42676f87e3e7d90f4599ae784d1edd3fb80d018cee38527da68ad0e5b08666d7"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.419576 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" podStartSLOduration=125.419563522 podStartE2EDuration="2m5.419563522s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.417403297 +0000 UTC m=+147.710054186" watchObservedRunningTime="2025-10-04 10:52:50.419563522 +0000 UTC m=+147.712214401" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.437673 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" event={"ID":"050b26f7-6ca2-4222-b2e0-fa32f256f234","Type":"ContainerStarted","Data":"ffd4ba7b7847c9476edb65977f6286c68a01abcb6d3ec6a798ea47c95ba40205"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.437715 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" event={"ID":"050b26f7-6ca2-4222-b2e0-fa32f256f234","Type":"ContainerStarted","Data":"bdb8725f7b85849d464b5fe9feae4f989f1b24b779ed58bf168e34a67483a404"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.448141 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" event={"ID":"583b0f46-ae37-4ba6-9d33-81fd3b22f633","Type":"ContainerStarted","Data":"0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.448172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" event={"ID":"583b0f46-ae37-4ba6-9d33-81fd3b22f633","Type":"ContainerStarted","Data":"b6896109eee3c66b695e19118c39ed882471cbc8942ff73cd10c613b494e40d7"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.449136 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.450226 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-s4v4x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.450276 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.460754 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.461503 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:50.96148782 +0000 UTC m=+148.254138709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.516690 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" event={"ID":"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b","Type":"ContainerStarted","Data":"0d9d7677f7120c7fc55484e3b9339d7a5831a7a179839e31d804b15dcc47658a"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.516733 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" event={"ID":"bf73a85e-8fab-45b4-ad46-4bf96bd00c1b","Type":"ContainerStarted","Data":"64075f2fd41efaa0b0a48daeab575dc4d5432c3d9b33386451af90cab0130d86"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.518683 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-t6jc6" event={"ID":"9c1109b9-7685-4315-a3b2-3f3f0d31f861","Type":"ContainerStarted","Data":"7b40670c2113e7405bdcd7cbb320e189d540516e4d56d3c20c24ac036a883707"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.544697 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:50 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:50 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:50 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.544976 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.549856 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" event={"ID":"5809e939-725d-48cb-a582-e115f63873ad","Type":"ContainerStarted","Data":"ec25281482225e4b2940f852842becf6e86deb3deffc9bb71b0ae0f81bd9c3e7"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.566503 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" event={"ID":"a7c28209-920e-49c3-9b2b-334dc06df888","Type":"ContainerStarted","Data":"cb4935e21eae4fc56d4436b2fedb3882cb8b772254b3e8e01904491898c7e8df"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.571606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.572585 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.072570505 +0000 UTC m=+148.365221394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.612260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" event={"ID":"e6ba4505-f064-4b1d-a545-25e3179485eb","Type":"ContainerStarted","Data":"3405b32a2de15dfb736caeaaddbbafcf724a380c9e355ba877c33a2704c189c0"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.612298 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" event={"ID":"e6ba4505-f064-4b1d-a545-25e3179485eb","Type":"ContainerStarted","Data":"02664b04256067061faec15496ceb2ee0f8ad4623f8ff7fc92782bf214a7a060"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.639031 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" event={"ID":"003701dd-2cdf-4ac6-8835-9017cc03039c","Type":"ContainerStarted","Data":"addbdb974bef8df9c088de6aed1bd3cc3918b3c341fd4fbf574ff846bbd02b7b"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.642465 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" podStartSLOduration=125.64245517 podStartE2EDuration="2m5.64245517s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.630394365 +0000 UTC m=+147.923045244" watchObservedRunningTime="2025-10-04 10:52:50.64245517 +0000 UTC m=+147.935106059" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.642966 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" podStartSLOduration=125.642959853 podStartE2EDuration="2m5.642959853s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.492553315 +0000 UTC m=+147.785204204" watchObservedRunningTime="2025-10-04 10:52:50.642959853 +0000 UTC m=+147.935610742" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.652258 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" event={"ID":"62cf7efa-32a9-4564-875e-b7b6b619805f","Type":"ContainerStarted","Data":"c44e6169ec2377f1a10148d7266677ce2253a7b5158ed86159643de588600d78"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.665617 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" event={"ID":"4c65edd1-18dd-4b95-81c5-c0a4e5d6013d","Type":"ContainerStarted","Data":"796230301d08530b7d377bb8bbe445c49d70e904af93ca7d272e6d84365a188f"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.667764 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" event={"ID":"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22","Type":"ContainerStarted","Data":"d6406a0f065df1b15c81f98d35602fef0aa0deeb68240efce327ee99527f3152"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.667784 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" event={"ID":"37dc3e36-9baf-4e27-9fc0-ddbe4c25dc22","Type":"ContainerStarted","Data":"5b38785d3e4718adb81a0f081438a960b3e5814bf4bb0c5506c14599ddc5aa64"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.668132 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.674729 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.676032 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.176020168 +0000 UTC m=+148.468671047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.679343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-brblh" event={"ID":"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5","Type":"ContainerStarted","Data":"5c8974ef1cdd940e11d20d7d05b1f342809fdd8e8db7d7edb298bd218a0bc033"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.682918 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" event={"ID":"957b0629-df4e-473b-9c09-eb9430312c69","Type":"ContainerStarted","Data":"9dd159c11eadb75f9e926bc137b7c65cce022a483e14628d0d97bd9812d65441"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.688450 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" event={"ID":"1fe83ad0-c0fa-4b95-9fbf-111dc0b6df0b","Type":"ContainerStarted","Data":"88144284bc0864b392179f1970d53a75575c89ba223702760e6b5f8a9d504f60"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.697353 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" event={"ID":"ce18ba60-323b-47c5-8230-81e47fb74a72","Type":"ContainerStarted","Data":"2c75f2b23aa19b7322790599433f863c614d1f427af666310a7ce21acdef3398"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.699131 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" event={"ID":"23d17080-6640-4c42-8c07-90b6fe9fa69f","Type":"ContainerStarted","Data":"f951bb46edf2fd4cccdbdceae6c8f8a2e86fd8add3ea850fb52dc20bfc6e0e70"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.699152 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" event={"ID":"23d17080-6640-4c42-8c07-90b6fe9fa69f","Type":"ContainerStarted","Data":"0842f155a2fe56224fee7bf2fcac3732c01b1f003b7ee76f14e72a8430997351"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.710587 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pqnhz" event={"ID":"aed3108c-fee1-4f60-b31f-ba579cbb34b6","Type":"ContainerStarted","Data":"f061625caf05c1e25ab4e04180da4e4ab9cf9c5bd9ed7fb1d57daf01aeb0c964"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.712004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jf9xm" event={"ID":"11fe31d1-ad44-4482-ba21-7c0c8a4d8a03","Type":"ContainerStarted","Data":"de5c9c15b9c87427b58eaaf80e59c39a701d8eda552eaef7033a24b238aa252d"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.713264 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" event={"ID":"78a37133-e23f-4bb2-a42d-b55800e168ca","Type":"ContainerStarted","Data":"78784ae4b720a9cdc857a1b5769889ac53a9b572ae07373fe0026dd86e780365"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.714645 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" event={"ID":"0b57bdb3-1a08-4a62-b4c2-ee25c470f84d","Type":"ContainerStarted","Data":"254f4f6febfc758ceb4eed203564d950a962f66db7c386086799cdb433ae39a0"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.716960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5ff4f" event={"ID":"9437fcf9-6541-495a-a30c-49c409b40155","Type":"ContainerStarted","Data":"cd55a2920cb700aaa4973b81d161c1ae8fb62c08d70240ab5b3660ed5ec06b11"} Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.730718 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lpsnq" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.778169 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.779358 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.279344456 +0000 UTC m=+148.571995345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.794801 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.844635 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-t6jc6" podStartSLOduration=7.844593394 podStartE2EDuration="7.844593394s" podCreationTimestamp="2025-10-04 10:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.756643923 +0000 UTC m=+148.049294812" watchObservedRunningTime="2025-10-04 10:52:50.844593394 +0000 UTC m=+148.137244283" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.881607 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.882046 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.382033489 +0000 UTC m=+148.674684378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.903785 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mhw4" podStartSLOduration=125.903767888 podStartE2EDuration="2m5.903767888s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.899753167 +0000 UTC m=+148.192404066" watchObservedRunningTime="2025-10-04 10:52:50.903767888 +0000 UTC m=+148.196418767" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.905277 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" podStartSLOduration=125.905271506 podStartE2EDuration="2m5.905271506s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.84444061 +0000 UTC m=+148.137091499" watchObservedRunningTime="2025-10-04 10:52:50.905271506 +0000 UTC m=+148.197922395" Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.990221 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:50 crc kubenswrapper[4758]: E1004 10:52:50.990914 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.490894368 +0000 UTC m=+148.783545257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:50 crc kubenswrapper[4758]: I1004 10:52:50.991142 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w48jm" podStartSLOduration=125.991093333 podStartE2EDuration="2m5.991093333s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:50.98304411 +0000 UTC m=+148.275694999" watchObservedRunningTime="2025-10-04 10:52:50.991093333 +0000 UTC m=+148.283744232" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.092232 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.092510 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.592498314 +0000 UTC m=+148.885149203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.140567 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" podStartSLOduration=126.140551937 podStartE2EDuration="2m6.140551937s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.102280351 +0000 UTC m=+148.394931240" watchObservedRunningTime="2025-10-04 10:52:51.140551937 +0000 UTC m=+148.433202826" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.192884 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.193035 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.193090 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.193289 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.693265638 +0000 UTC m=+148.985916527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.194553 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.211047 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.231607 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-pqnhz" podStartSLOduration=126.231589206 podStartE2EDuration="2m6.231589206s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.220632779 +0000 UTC m=+148.513283668" watchObservedRunningTime="2025-10-04 10:52:51.231589206 +0000 UTC m=+148.524240095" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.232258 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j2rqz" podStartSLOduration=126.232252433 podStartE2EDuration="2m6.232252433s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.140155587 +0000 UTC m=+148.432806476" watchObservedRunningTime="2025-10-04 10:52:51.232252433 +0000 UTC m=+148.524903322" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.294513 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.294583 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.294623 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.294928 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.794912315 +0000 UTC m=+149.087563204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.298532 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.298843 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.350490 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.368121 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.396173 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.396404 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.896381147 +0000 UTC m=+149.189032036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.396496 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.396769 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.896762457 +0000 UTC m=+149.189413346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.402913 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.497468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.497713 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:51.997698986 +0000 UTC m=+149.290349875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.540296 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:51 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:51 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:51 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.540370 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.547013 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gpsj2" podStartSLOduration=126.54700032 podStartE2EDuration="2m6.54700032s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.356977212 +0000 UTC m=+148.649628101" watchObservedRunningTime="2025-10-04 10:52:51.54700032 +0000 UTC m=+148.839651199" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.598816 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.599075 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.099064035 +0000 UTC m=+149.391714924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.601893 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4wrsp" podStartSLOduration=126.601874476 podStartE2EDuration="2m6.601874476s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.594465689 +0000 UTC m=+148.887116578" watchObservedRunningTime="2025-10-04 10:52:51.601874476 +0000 UTC m=+148.894525355" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.603220 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" podStartSLOduration=126.60321529 podStartE2EDuration="2m6.60321529s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.549911244 +0000 UTC m=+148.842562133" watchObservedRunningTime="2025-10-04 10:52:51.60321529 +0000 UTC m=+148.895866179" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.701383 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.701924 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.201906692 +0000 UTC m=+149.494557581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.719208 4758 patch_prober.go:28] interesting pod/console-operator-58897d9998-sdqlk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.719267 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" podUID="1f8d7281-c6e3-4614-978e-ae5d4c170a85" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.733818 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" event={"ID":"c83e6e7d-1ec6-4e99-aa5b-c7490a86bfa1","Type":"ContainerStarted","Data":"849865ef1f4acdc435ef4649128c54b5ef5e1fd2fa3d61214d98324fdbc384b8"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.735670 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pqnhz" event={"ID":"aed3108c-fee1-4f60-b31f-ba579cbb34b6","Type":"ContainerStarted","Data":"8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.747686 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" event={"ID":"050b26f7-6ca2-4222-b2e0-fa32f256f234","Type":"ContainerStarted","Data":"b60f4a9df1c300a4035b614c5b5e82c20f439bddedb4be9395a2c28b6fd70dbe"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.763558 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" event={"ID":"67d42404-ab13-451a-84a1-4aba57e7d8bb","Type":"ContainerStarted","Data":"2ff9e353a0a476f9c5f33bc55a2f77a21c60c88ad2529093d4f1e28c22270766"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.766240 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" event={"ID":"907568c8-1873-41b3-8a58-3651ee7e7922","Type":"ContainerStarted","Data":"a1ae3981b7cfd72a2518a5832d3bf2fb9beead73992b8c9683c40a3aa285fc9f"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.767643 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" event={"ID":"ac7d66c4-cd59-4646-8970-5da893df6a56","Type":"ContainerStarted","Data":"1aa7223945b4d88c107c60e1827c7e53b000a22d4f8ee0d699f073dc2438f10b"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.769140 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" event={"ID":"5809e939-725d-48cb-a582-e115f63873ad","Type":"ContainerStarted","Data":"a2ffd27a1f7d3a5caf81921319a7f663bcdd6eb8c552fb8bd261ff6baa92496c"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.807804 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.810061 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.310050723 +0000 UTC m=+149.602701612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.812796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" event={"ID":"957b0629-df4e-473b-9c09-eb9430312c69","Type":"ContainerStarted","Data":"7e4550ed0f8585d35c44825a5b44bf76b8f645f06a262b94c4f45ccd5842a4e7"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.815566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-brblh" event={"ID":"fc64d8cc-0cd0-43cc-b0d6-89099fc3a0b5","Type":"ContainerStarted","Data":"47975cb644f578633bb4385a7080e78eec1ac43931035b3c29c627c26d5a720b"} Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.815595 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-brblh" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.818297 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-s4v4x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.818345 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.837975 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vdnpz" Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.908913 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.909254 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.409124294 +0000 UTC m=+149.701775183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:51 crc kubenswrapper[4758]: I1004 10:52:51.909918 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:51 crc kubenswrapper[4758]: E1004 10:52:51.922436 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.4224222 +0000 UTC m=+149.715073089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.011682 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.011981 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.511966401 +0000 UTC m=+149.804617290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.030723 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qq482" podStartSLOduration=127.030708744 podStartE2EDuration="2m7.030708744s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:51.831229137 +0000 UTC m=+149.123880026" watchObservedRunningTime="2025-10-04 10:52:52.030708744 +0000 UTC m=+149.323359633" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.117892 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.118315 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.618300216 +0000 UTC m=+149.910951105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.220258 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.220870 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.720856065 +0000 UTC m=+150.013506954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.312762 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm4jl" podStartSLOduration=127.312747105 podStartE2EDuration="2m7.312747105s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:52.053402707 +0000 UTC m=+149.346053596" watchObservedRunningTime="2025-10-04 10:52:52.312747105 +0000 UTC m=+149.605397984" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.322069 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.322343 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.822332947 +0000 UTC m=+150.114983836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.413141 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9k96q" podStartSLOduration=127.41312625 podStartE2EDuration="2m7.41312625s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:52.313214057 +0000 UTC m=+149.605864946" watchObservedRunningTime="2025-10-04 10:52:52.41312625 +0000 UTC m=+149.705777129" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.423448 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.423718 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:52.923703117 +0000 UTC m=+150.216354006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: W1004 10:52:52.438775 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-106fcba85423ae616353331ce086c4fad74abed0973654b8486eac444c4f297c WatchSource:0}: Error finding container 106fcba85423ae616353331ce086c4fad74abed0973654b8486eac444c4f297c: Status 404 returned error can't find the container with id 106fcba85423ae616353331ce086c4fad74abed0973654b8486eac444c4f297c Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.525430 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.526025 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.02601332 +0000 UTC m=+150.318664209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.540667 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:52 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:52 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:52 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.540724 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.563922 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4vc25" podStartSLOduration=127.563906617 podStartE2EDuration="2m7.563906617s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:52.562338548 +0000 UTC m=+149.854989437" watchObservedRunningTime="2025-10-04 10:52:52.563906617 +0000 UTC m=+149.856557506" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.630152 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.630401 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.130386766 +0000 UTC m=+150.423037645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.639787 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4z96l" podStartSLOduration=127.639774923 podStartE2EDuration="2m7.639774923s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:52.621014939 +0000 UTC m=+149.913665828" watchObservedRunningTime="2025-10-04 10:52:52.639774923 +0000 UTC m=+149.932425812" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.703199 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4h92h" Oct 04 10:52:52 crc kubenswrapper[4758]: W1004 10:52:52.721573 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-fe992c6629b0d79b03c2fdbf01687060943c44fc33649bfd33dbc7c2ab2001fa WatchSource:0}: Error finding container fe992c6629b0d79b03c2fdbf01687060943c44fc33649bfd33dbc7c2ab2001fa: Status 404 returned error can't find the container with id fe992c6629b0d79b03c2fdbf01687060943c44fc33649bfd33dbc7c2ab2001fa Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.732663 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.732941 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.232928355 +0000 UTC m=+150.525579244 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.816551 4758 patch_prober.go:28] interesting pod/console-operator-58897d9998-sdqlk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.816596 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" podUID="1f8d7281-c6e3-4614-978e-ae5d4c170a85" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.818522 4758 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tfqb6 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.818557 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" podUID="f4c724dc-0661-4145-b264-eb28df91be33" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.833680 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.833978 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.333963726 +0000 UTC m=+150.626614615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.855978 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fe992c6629b0d79b03c2fdbf01687060943c44fc33649bfd33dbc7c2ab2001fa"} Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.858075 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" event={"ID":"ac7d66c4-cd59-4646-8970-5da893df6a56","Type":"ContainerStarted","Data":"50e53cb7a6f5d5f34590a50d89fb824fb558d66932789f7a5dc7ef9b91aa9ccb"} Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.858906 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"015339303b94203a81ecf8537104a0c4b66417e69390f9d314e7edde59129183"} Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.858930 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"106fcba85423ae616353331ce086c4fad74abed0973654b8486eac444c4f297c"} Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.860466 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c608a062c35c14f98a291243150d12898892e1a148fe8cdee88e5a19690cdcda"} Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.862129 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-s4v4x container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.862192 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.936262 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:52 crc kubenswrapper[4758]: E1004 10:52:52.936627 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.436616728 +0000 UTC m=+150.729267607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:52 crc kubenswrapper[4758]: I1004 10:52:52.979493 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2vd2f" podStartSLOduration=127.979478261 podStartE2EDuration="2m7.979478261s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:52.919385093 +0000 UTC m=+150.212035982" watchObservedRunningTime="2025-10-04 10:52:52.979478261 +0000 UTC m=+150.272129150" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.036773 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.038089 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.53807377 +0000 UTC m=+150.830724659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.109641 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-t9c2p" podStartSLOduration=128.109625217 podStartE2EDuration="2m8.109625217s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:53.101295007 +0000 UTC m=+150.393945896" watchObservedRunningTime="2025-10-04 10:52:53.109625217 +0000 UTC m=+150.402276106" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.112559 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-l6frd" podStartSLOduration=128.112550751 podStartE2EDuration="2m8.112550751s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:53.067442662 +0000 UTC m=+150.360093551" watchObservedRunningTime="2025-10-04 10:52:53.112550751 +0000 UTC m=+150.405201630" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.140652 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.141213 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.641200564 +0000 UTC m=+150.933851453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.206266 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z5ck8"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.207133 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.223441 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.241456 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.241705 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.741690262 +0000 UTC m=+151.034341151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.242919 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" podStartSLOduration=128.242908382 podStartE2EDuration="2m8.242908382s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:53.213088829 +0000 UTC m=+150.505739718" watchObservedRunningTime="2025-10-04 10:52:53.242908382 +0000 UTC m=+150.535559271" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.248996 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5ck8"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.283193 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-brblh" podStartSLOduration=10.283177559 podStartE2EDuration="10.283177559s" podCreationTimestamp="2025-10-04 10:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:53.282442811 +0000 UTC m=+150.575093700" watchObservedRunningTime="2025-10-04 10:52:53.283177559 +0000 UTC m=+150.575828448" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.342793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86m7q\" (UniqueName: \"kubernetes.io/projected/3bf4f130-658a-40c0-97c5-243d77950128-kube-api-access-86m7q\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.342862 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.342908 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-utilities\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.342927 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-catalog-content\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.343199 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.843189685 +0000 UTC m=+151.135840574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.350443 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-n6j98" podStartSLOduration=128.350429127 podStartE2EDuration="2m8.350429127s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:53.349967116 +0000 UTC m=+150.642618005" watchObservedRunningTime="2025-10-04 10:52:53.350429127 +0000 UTC m=+150.643080026" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.378340 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6xncn"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.389643 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.409576 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.413016 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6xncn"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.444683 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.445158 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.945141069 +0000 UTC m=+151.237791958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.445267 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.445407 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-utilities\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.445511 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-catalog-content\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.445612 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86m7q\" (UniqueName: \"kubernetes.io/projected/3bf4f130-658a-40c0-97c5-243d77950128-kube-api-access-86m7q\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.446342 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:53.946333579 +0000 UTC m=+151.238984458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.447031 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-utilities\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.447404 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-catalog-content\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.511955 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86m7q\" (UniqueName: \"kubernetes.io/projected/3bf4f130-658a-40c0-97c5-243d77950128-kube-api-access-86m7q\") pod \"certified-operators-z5ck8\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.554435 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.555227 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.555475 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-catalog-content\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.555588 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.055565747 +0000 UTC m=+151.348216626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.555697 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5t6\" (UniqueName: \"kubernetes.io/projected/447f9eb2-901e-42c0-813c-f03ac5f02916-kube-api-access-vn5t6\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.555806 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.555903 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-utilities\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.556282 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.056270135 +0000 UTC m=+151.348921024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.581288 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:53 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:53 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:53 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.582424 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.591371 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bfvsd"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.614805 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.632275 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bfvsd"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.664462 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.664593 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.16457421 +0000 UTC m=+151.457225099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.664679 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-catalog-content\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.664706 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5t6\" (UniqueName: \"kubernetes.io/projected/447f9eb2-901e-42c0-813c-f03ac5f02916-kube-api-access-vn5t6\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.664741 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.664766 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-utilities\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.665380 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-utilities\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.665746 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.165708908 +0000 UTC m=+151.458359797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.666202 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-catalog-content\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.726023 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5t6\" (UniqueName: \"kubernetes.io/projected/447f9eb2-901e-42c0-813c-f03ac5f02916-kube-api-access-vn5t6\") pod \"community-operators-6xncn\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.765824 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.766031 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-utilities\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.766092 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2zbx\" (UniqueName: \"kubernetes.io/projected/9c3a39ce-478e-4c93-b415-03700dcb90e1-kube-api-access-f2zbx\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.766142 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-catalog-content\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.766262 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.266246057 +0000 UTC m=+151.558896946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.853398 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2pw5r"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.855168 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.868124 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2zbx\" (UniqueName: \"kubernetes.io/projected/9c3a39ce-478e-4c93-b415-03700dcb90e1-kube-api-access-f2zbx\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.868171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.868205 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-catalog-content\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.868245 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-utilities\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.868862 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.368848238 +0000 UTC m=+151.661499127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.868936 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-utilities\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.869258 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-catalog-content\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.880644 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2pw5r"] Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.897491 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" event={"ID":"ac7d66c4-cd59-4646-8970-5da893df6a56","Type":"ContainerStarted","Data":"2e8f4f68932a8a416d7df6168d0b1adf9b91731d00e55ebf1aaa0ba1f0463a85"} Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.898833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1dfbb7048785e6fffd229114b331499cedf17614206997cba8003f7a1f28a231"} Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.898837 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2zbx\" (UniqueName: \"kubernetes.io/projected/9c3a39ce-478e-4c93-b415-03700dcb90e1-kube-api-access-f2zbx\") pod \"certified-operators-bfvsd\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.899397 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.900574 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"59c7cf0ebd79679601ea6805ee6d673be59b1cb914f2b99063c5f5b9e3871e17"} Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.971160 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.971403 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-utilities\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.971452 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wcc\" (UniqueName: \"kubernetes.io/projected/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-kube-api-access-b5wcc\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.971485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-catalog-content\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:53 crc kubenswrapper[4758]: E1004 10:52:53.971616 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.471600132 +0000 UTC m=+151.764251021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:53 crc kubenswrapper[4758]: I1004 10:52:53.981358 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.025421 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.074178 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-catalog-content\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.074230 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.074306 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-utilities\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.074363 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5wcc\" (UniqueName: \"kubernetes.io/projected/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-kube-api-access-b5wcc\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.075819 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-catalog-content\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.076041 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.576029409 +0000 UTC m=+151.868680298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.076650 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-utilities\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.152068 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5wcc\" (UniqueName: \"kubernetes.io/projected/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-kube-api-access-b5wcc\") pod \"community-operators-2pw5r\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.175089 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.175330 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.675315836 +0000 UTC m=+151.967966725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.183480 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.276080 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.276404 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.776393519 +0000 UTC m=+152.069044408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.377697 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.377967 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.877952503 +0000 UTC m=+152.170603392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.480976 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.481428 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:54.981412206 +0000 UTC m=+152.274063095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.539925 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:54 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:54 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:54 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.539983 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.546473 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5ck8"] Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.583291 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.583423 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.083394891 +0000 UTC m=+152.376045780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.583474 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.583761 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.08374921 +0000 UTC m=+152.376400099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.686359 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.686954 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.186939085 +0000 UTC m=+152.479589974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.791367 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.791739 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.291723081 +0000 UTC m=+152.584373970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.801175 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.801773 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.810402 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.813274 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.873611 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.892909 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.893146 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.393117971 +0000 UTC m=+152.685768860 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.893189 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.893238 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.893281 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.893810 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.393802129 +0000 UTC m=+152.686453018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.907383 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerStarted","Data":"3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0"} Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.907446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerStarted","Data":"87c5ef0ad85aa3f4757bb167dfedde0c9a90d0453433737cde2ecf42f3358f0a"} Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.910796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" event={"ID":"ac7d66c4-cd59-4646-8970-5da893df6a56","Type":"ContainerStarted","Data":"ffeafd1e38f542b2f784cfd05c1d69d089cd15afd5214c3b2a645bef595f0bf2"} Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.994371 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.994869 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.994937 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:54 crc kubenswrapper[4758]: I1004 10:52:54.995089 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:54 crc kubenswrapper[4758]: E1004 10:52:54.995216 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.495197809 +0000 UTC m=+152.787848698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.071993 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.097707 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.098034 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.598022495 +0000 UTC m=+152.890673384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.114619 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.198521 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.198842 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.698825951 +0000 UTC m=+152.991476840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.255261 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-4cxfd" podStartSLOduration=12.255245165 podStartE2EDuration="12.255245165s" podCreationTimestamp="2025-10-04 10:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:55.086647538 +0000 UTC m=+152.379298427" watchObservedRunningTime="2025-10-04 10:52:55.255245165 +0000 UTC m=+152.547896054" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.255594 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2pw5r"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.299801 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.300121 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.800090718 +0000 UTC m=+153.092741607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.307089 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bfvsd"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.360084 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6xncn"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.396051 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xmk5z"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.397013 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.401434 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.402486 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.402774 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:55.90275851 +0000 UTC m=+153.195409389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.435083 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xmk5z"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.504708 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qkmd\" (UniqueName: \"kubernetes.io/projected/096babec-37b1-4321-a5f6-b3a614a0b49e-kube-api-access-4qkmd\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.504765 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-utilities\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.504782 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-catalog-content\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.504827 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.505079 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.005068304 +0000 UTC m=+153.297719193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.548303 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:55 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:55 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:55 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.548349 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.563476 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-dwm2t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.563532 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dwm2t" podUID="dfcb1f56-7b71-4cf6-a92d-8f174194e697" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.563592 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-dwm2t container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.563605 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dwm2t" podUID="dfcb1f56-7b71-4cf6-a92d-8f174194e697" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.610584 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.610975 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-utilities\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.610997 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-catalog-content\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.611060 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qkmd\" (UniqueName: \"kubernetes.io/projected/096babec-37b1-4321-a5f6-b3a614a0b49e-kube-api-access-4qkmd\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.611759 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-catalog-content\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.611812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-utilities\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.611822 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.111808289 +0000 UTC m=+153.404459168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.625253 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.630227 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.648286 4758 patch_prober.go:28] interesting pod/apiserver-76f77b778f-gjvgj container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]log ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]etcd ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/max-in-flight-filter ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 04 10:52:55 crc kubenswrapper[4758]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/project.openshift.io-projectcache ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/openshift.io-startinformers ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 04 10:52:55 crc kubenswrapper[4758]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 10:52:55 crc kubenswrapper[4758]: livez check failed Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.648333 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" podUID="957b0629-df4e-473b-9c09-eb9430312c69" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.674838 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qkmd\" (UniqueName: \"kubernetes.io/projected/096babec-37b1-4321-a5f6-b3a614a0b49e-kube-api-access-4qkmd\") pod \"redhat-marketplace-xmk5z\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.716843 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.718194 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.718490 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.218474932 +0000 UTC m=+153.511125821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.789305 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c8zz9"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.790177 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.799866 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.800247 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.811173 4758 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.813480 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8zz9"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.831588 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.831920 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.331905475 +0000 UTC m=+153.624556364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.835643 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.929859 4758 generic.go:334] "Generic (PLEG): container finished" podID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerID="dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e" exitCode=0 Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.929999 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfvsd" event={"ID":"9c3a39ce-478e-4c93-b415-03700dcb90e1","Type":"ContainerDied","Data":"dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e"} Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.930051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfvsd" event={"ID":"9c3a39ce-478e-4c93-b415-03700dcb90e1","Type":"ContainerStarted","Data":"78d66a881117e1dc76677e429ec488379ca3a22cede5ce15582ae34b05d16895"} Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.931702 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.932596 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-catalog-content\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.932829 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.932979 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzmtj\" (UniqueName: \"kubernetes.io/projected/45fc0b19-de79-477d-b2f4-db9786113587-kube-api-access-kzmtj\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.932999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-utilities\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:55 crc kubenswrapper[4758]: E1004 10:52:55.933685 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.433667805 +0000 UTC m=+153.726318694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.934374 4758 generic.go:334] "Generic (PLEG): container finished" podID="3bf4f130-658a-40c0-97c5-243d77950128" containerID="3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0" exitCode=0 Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.934460 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerDied","Data":"3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0"} Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.954227 4758 generic.go:334] "Generic (PLEG): container finished" podID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerID="cb76b8fae751853dfdaf60c47c8b813c1315c3bc332d24e5a86222e3f998cc6f" exitCode=0 Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.955029 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xncn" event={"ID":"447f9eb2-901e-42c0-813c-f03ac5f02916","Type":"ContainerDied","Data":"cb76b8fae751853dfdaf60c47c8b813c1315c3bc332d24e5a86222e3f998cc6f"} Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.955057 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xncn" event={"ID":"447f9eb2-901e-42c0-813c-f03ac5f02916","Type":"ContainerStarted","Data":"967a467e90f214e5e95f7d5dc123bc5dd172b9710505d22bc135e51432923c0d"} Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.974489 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.974670 4758 generic.go:334] "Generic (PLEG): container finished" podID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerID="abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96" exitCode=0 Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.976344 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pw5r" event={"ID":"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e","Type":"ContainerDied","Data":"abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96"} Oct 04 10:52:55 crc kubenswrapper[4758]: I1004 10:52:55.976366 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pw5r" event={"ID":"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e","Type":"ContainerStarted","Data":"37da6e968180506a82ad20d3eaa7310e1e4fc23e352b48cbb34fd0c8491998bf"} Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.000770 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rgz5g" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.036455 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sdqlk" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.037310 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.037532 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-utilities\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.037579 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-catalog-content\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.037670 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzmtj\" (UniqueName: \"kubernetes.io/projected/45fc0b19-de79-477d-b2f4-db9786113587-kube-api-access-kzmtj\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.037945 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.537931458 +0000 UTC m=+153.830582347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.038323 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-utilities\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.039444 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-catalog-content\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.095601 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzmtj\" (UniqueName: \"kubernetes.io/projected/45fc0b19-de79-477d-b2f4-db9786113587-kube-api-access-kzmtj\") pod \"redhat-marketplace-c8zz9\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.136259 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.146251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.150145 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.650128251 +0000 UTC m=+153.942779140 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.248765 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.249250 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.749235773 +0000 UTC m=+154.041886662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.350639 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.350940 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.850926881 +0000 UTC m=+154.143577760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.411633 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qgnlj"] Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.412893 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: W1004 10:52:56.419497 4758 reflector.go:561] object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh": failed to list *v1.Secret: secrets "redhat-operators-dockercfg-ct8rh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.419537 4758 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-ct8rh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-operators-dockercfg-ct8rh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.456326 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.456523 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-catalog-content\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.456581 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lprs9\" (UniqueName: \"kubernetes.io/projected/688d59c7-04be-4f21-ae68-f134a6b72e66-kube-api-access-lprs9\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.456599 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-utilities\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.456710 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:56.956694042 +0000 UTC m=+154.249344931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.543360 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qgnlj"] Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.544688 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.549275 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:56 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:56 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:56 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.549326 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.550230 4758 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T10:52:55.811190472Z","Handler":null,"Name":""} Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.557207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-catalog-content\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.557259 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.557306 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lprs9\" (UniqueName: \"kubernetes.io/projected/688d59c7-04be-4f21-ae68-f134a6b72e66-kube-api-access-lprs9\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.557322 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-utilities\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.557702 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-utilities\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.558956 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-catalog-content\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.559356 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:52:57.059344694 +0000 UTC m=+154.351995583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gr62g" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.598780 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lprs9\" (UniqueName: \"kubernetes.io/projected/688d59c7-04be-4f21-ae68-f134a6b72e66-kube-api-access-lprs9\") pod \"redhat-operators-qgnlj\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.657870 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:56 crc kubenswrapper[4758]: E1004 10:52:56.658119 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:52:57.158091797 +0000 UTC m=+154.450742686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.734501 4758 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.734538 4758 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.768147 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.778225 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rnp22"] Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.779158 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.781260 4758 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.781320 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.806639 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnp22"] Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.827607 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xmk5z"] Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.948388 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8zz9"] Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.960466 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.973267 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52744\" (UniqueName: \"kubernetes.io/projected/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-kube-api-access-52744\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.973355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-catalog-content\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.973395 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-utilities\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:56 crc kubenswrapper[4758]: I1004 10:52:56.975008 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tfqb6" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.049158 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xmk5z" event={"ID":"096babec-37b1-4321-a5f6-b3a614a0b49e","Type":"ContainerStarted","Data":"9d0e04431c4966e94ddbc0d92fd1ac9b37efe7989c977b20f0d36b112519d9c6"} Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.066005 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gr62g\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.066372 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50","Type":"ContainerStarted","Data":"281befe8661d5116223f94a807b5dcf1da82e22d4d471b6a3722418092b3b18f"} Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.066405 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50","Type":"ContainerStarted","Data":"2a6fc708b2df2a6150cf1258afc20a4dc9aa56615a8ec0a9d268f0a2274ecb70"} Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.077923 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.078060 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-catalog-content\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.078128 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-utilities\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.078187 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52744\" (UniqueName: \"kubernetes.io/projected/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-kube-api-access-52744\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.079416 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-catalog-content\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.096264 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-utilities\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.096962 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.096952259 podStartE2EDuration="3.096952259s" podCreationTimestamp="2025-10-04 10:52:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:57.09262334 +0000 UTC m=+154.385274239" watchObservedRunningTime="2025-10-04 10:52:57.096952259 +0000 UTC m=+154.389603148" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.098076 4758 generic.go:334] "Generic (PLEG): container finished" podID="a7c28209-920e-49c3-9b2b-334dc06df888" containerID="cb4935e21eae4fc56d4436b2fedb3882cb8b772254b3e8e01904491898c7e8df" exitCode=0 Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.098737 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" event={"ID":"a7c28209-920e-49c3-9b2b-334dc06df888","Type":"ContainerDied","Data":"cb4935e21eae4fc56d4436b2fedb3882cb8b772254b3e8e01904491898c7e8df"} Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.111815 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52744\" (UniqueName: \"kubernetes.io/projected/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-kube-api-access-52744\") pod \"redhat-operators-rnp22\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.167680 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.167716 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.171260 4758 patch_prober.go:28] interesting pod/console-f9d7485db-pqnhz container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.171306 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pqnhz" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.221575 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.322741 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.340606 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.515633 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gr62g"] Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.541019 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:57 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:57 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:57 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.541117 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.760598 4758 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/redhat-operators-qgnlj" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.760922 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.765867 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 10:52:57 crc kubenswrapper[4758]: I1004 10:52:57.774530 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.068751 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnp22"] Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.138320 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" event={"ID":"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a","Type":"ContainerStarted","Data":"a11e8b59ab8f3465c28f2e6620e108ff0a95f1e40f413d8d4a93915c3fb43b44"} Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.139657 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnp22" event={"ID":"d3f8a403-d7ef-46d3-987e-f2d9b89f3172","Type":"ContainerStarted","Data":"6db9d2767070633c3f3793f845fa360a92daec9e79d34b764a277f7d6af33d1f"} Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.140942 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerStarted","Data":"87129a6c82e52a0acef61af70d9c34a9f3bc8d34eb3f3d9cf9a7ee98b4da7dfa"} Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.140960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerStarted","Data":"e9ce8a9562f20f7a18dfbec0bc8dd6acbdedda292ba277e70feb14c6d0010564"} Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.142465 4758 generic.go:334] "Generic (PLEG): container finished" podID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerID="ed457ff81c2c02f1393fb3263c4695afb40580b6dc2a415cce5c1ff1c7c18743" exitCode=0 Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.142509 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xmk5z" event={"ID":"096babec-37b1-4321-a5f6-b3a614a0b49e","Type":"ContainerDied","Data":"ed457ff81c2c02f1393fb3263c4695afb40580b6dc2a415cce5c1ff1c7c18743"} Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.150348 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qgnlj"] Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.155287 4758 generic.go:334] "Generic (PLEG): container finished" podID="b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50" containerID="281befe8661d5116223f94a807b5dcf1da82e22d4d471b6a3722418092b3b18f" exitCode=0 Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.156170 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50","Type":"ContainerDied","Data":"281befe8661d5116223f94a807b5dcf1da82e22d4d471b6a3722418092b3b18f"} Oct 04 10:52:58 crc kubenswrapper[4758]: W1004 10:52:58.189078 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod688d59c7_04be_4f21_ae68_f134a6b72e66.slice/crio-524d5a26e82daf35a537f324b141810575639bf7acb34ee54972fa7c5f39f1ae WatchSource:0}: Error finding container 524d5a26e82daf35a537f324b141810575639bf7acb34ee54972fa7c5f39f1ae: Status 404 returned error can't find the container with id 524d5a26e82daf35a537f324b141810575639bf7acb34ee54972fa7c5f39f1ae Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.430721 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.500076 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 10:52:58 crc kubenswrapper[4758]: E1004 10:52:58.500286 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c28209-920e-49c3-9b2b-334dc06df888" containerName="collect-profiles" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.500302 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c28209-920e-49c3-9b2b-334dc06df888" containerName="collect-profiles" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.500393 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7c28209-920e-49c3-9b2b-334dc06df888" containerName="collect-profiles" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.500731 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.503346 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.503527 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.506020 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.542325 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:58 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:58 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:58 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.542392 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.612667 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfl4q\" (UniqueName: \"kubernetes.io/projected/a7c28209-920e-49c3-9b2b-334dc06df888-kube-api-access-qfl4q\") pod \"a7c28209-920e-49c3-9b2b-334dc06df888\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.612760 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7c28209-920e-49c3-9b2b-334dc06df888-secret-volume\") pod \"a7c28209-920e-49c3-9b2b-334dc06df888\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.612821 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7c28209-920e-49c3-9b2b-334dc06df888-config-volume\") pod \"a7c28209-920e-49c3-9b2b-334dc06df888\" (UID: \"a7c28209-920e-49c3-9b2b-334dc06df888\") " Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.613004 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b06830a3-62da-4416-9621-65fced1b4073-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.613036 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b06830a3-62da-4416-9621-65fced1b4073-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.631442 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7c28209-920e-49c3-9b2b-334dc06df888-config-volume" (OuterVolumeSpecName: "config-volume") pod "a7c28209-920e-49c3-9b2b-334dc06df888" (UID: "a7c28209-920e-49c3-9b2b-334dc06df888"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.637554 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7c28209-920e-49c3-9b2b-334dc06df888-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a7c28209-920e-49c3-9b2b-334dc06df888" (UID: "a7c28209-920e-49c3-9b2b-334dc06df888"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.645262 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7c28209-920e-49c3-9b2b-334dc06df888-kube-api-access-qfl4q" (OuterVolumeSpecName: "kube-api-access-qfl4q") pod "a7c28209-920e-49c3-9b2b-334dc06df888" (UID: "a7c28209-920e-49c3-9b2b-334dc06df888"). InnerVolumeSpecName "kube-api-access-qfl4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.705346 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.714329 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b06830a3-62da-4416-9621-65fced1b4073-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.714371 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b06830a3-62da-4416-9621-65fced1b4073-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.714418 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7c28209-920e-49c3-9b2b-334dc06df888-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.714431 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfl4q\" (UniqueName: \"kubernetes.io/projected/a7c28209-920e-49c3-9b2b-334dc06df888-kube-api-access-qfl4q\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.714440 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7c28209-920e-49c3-9b2b-334dc06df888-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.714439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b06830a3-62da-4416-9621-65fced1b4073-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.741998 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b06830a3-62da-4416-9621-65fced1b4073-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:58 crc kubenswrapper[4758]: I1004 10:52:58.826298 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.171417 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" event={"ID":"a7c28209-920e-49c3-9b2b-334dc06df888","Type":"ContainerDied","Data":"007a0a0ea970e919e1b2573ca0af6a1fbd7eec539fe95843b58b0f2ca5070c6b"} Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.171472 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="007a0a0ea970e919e1b2573ca0af6a1fbd7eec539fe95843b58b0f2ca5070c6b" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.171441 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.173078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerStarted","Data":"cdc76eb5197605a1ad0d4bc26a3f8a7adec748a2b649eb15e146e970c02c8a26"} Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.173134 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerStarted","Data":"524d5a26e82daf35a537f324b141810575639bf7acb34ee54972fa7c5f39f1ae"} Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.176133 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnp22" event={"ID":"d3f8a403-d7ef-46d3-987e-f2d9b89f3172","Type":"ContainerDied","Data":"a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2"} Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.176091 4758 generic.go:334] "Generic (PLEG): container finished" podID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerID="a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2" exitCode=0 Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.178450 4758 generic.go:334] "Generic (PLEG): container finished" podID="45fc0b19-de79-477d-b2f4-db9786113587" containerID="87129a6c82e52a0acef61af70d9c34a9f3bc8d34eb3f3d9cf9a7ee98b4da7dfa" exitCode=0 Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.178808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerDied","Data":"87129a6c82e52a0acef61af70d9c34a9f3bc8d34eb3f3d9cf9a7ee98b4da7dfa"} Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.183313 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" event={"ID":"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a","Type":"ContainerStarted","Data":"7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99"} Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.185012 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.241181 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" podStartSLOduration=134.241162582 podStartE2EDuration="2m14.241162582s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:59.238203547 +0000 UTC m=+156.530854426" watchObservedRunningTime="2025-10-04 10:52:59.241162582 +0000 UTC m=+156.533813471" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.269711 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.540436 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:52:59 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:52:59 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:52:59 crc kubenswrapper[4758]: healthz check failed Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.540493 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.559767 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.730466 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kubelet-dir\") pod \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.730566 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50" (UID: "b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.730845 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kube-api-access\") pod \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\" (UID: \"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50\") " Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.731085 4758 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.736677 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50" (UID: "b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:59 crc kubenswrapper[4758]: I1004 10:52:59.831663 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.190905 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b06830a3-62da-4416-9621-65fced1b4073","Type":"ContainerStarted","Data":"375dd88c4a80c3b5eb360eb3d9311a39d53201cc477b16f558694470b5a9d6da"} Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.190958 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b06830a3-62da-4416-9621-65fced1b4073","Type":"ContainerStarted","Data":"0f1566bd0e9b8a26958b00007106856053b6fb8c3fed5806e8945d4f10b149f1"} Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.193193 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50","Type":"ContainerDied","Data":"2a6fc708b2df2a6150cf1258afc20a4dc9aa56615a8ec0a9d268f0a2274ecb70"} Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.193221 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a6fc708b2df2a6150cf1258afc20a4dc9aa56615a8ec0a9d268f0a2274ecb70" Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.193282 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.195023 4758 generic.go:334] "Generic (PLEG): container finished" podID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerID="cdc76eb5197605a1ad0d4bc26a3f8a7adec748a2b649eb15e146e970c02c8a26" exitCode=0 Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.195853 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerDied","Data":"cdc76eb5197605a1ad0d4bc26a3f8a7adec748a2b649eb15e146e970c02c8a26"} Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.217795 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.217774301 podStartE2EDuration="2.217774301s" podCreationTimestamp="2025-10-04 10:52:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:53:00.217067133 +0000 UTC m=+157.509718022" watchObservedRunningTime="2025-10-04 10:53:00.217774301 +0000 UTC m=+157.510425190" Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.539612 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:00 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:00 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:00 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.539678 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.630344 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:53:00 crc kubenswrapper[4758]: I1004 10:53:00.640403 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-gjvgj" Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.209083 4758 generic.go:334] "Generic (PLEG): container finished" podID="b06830a3-62da-4416-9621-65fced1b4073" containerID="375dd88c4a80c3b5eb360eb3d9311a39d53201cc477b16f558694470b5a9d6da" exitCode=0 Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.209982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b06830a3-62da-4416-9621-65fced1b4073","Type":"ContainerDied","Data":"375dd88c4a80c3b5eb360eb3d9311a39d53201cc477b16f558694470b5a9d6da"} Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.249203 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.249268 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.539968 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:01 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:01 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:01 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.540019 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:01 crc kubenswrapper[4758]: I1004 10:53:01.720665 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-brblh" Oct 04 10:53:02 crc kubenswrapper[4758]: I1004 10:53:02.540203 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:02 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:02 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:02 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:02 crc kubenswrapper[4758]: I1004 10:53:02.540277 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:03 crc kubenswrapper[4758]: I1004 10:53:03.539866 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:03 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:03 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:03 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:03 crc kubenswrapper[4758]: I1004 10:53:03.540150 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:04 crc kubenswrapper[4758]: I1004 10:53:04.539190 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:04 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:04 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:04 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:04 crc kubenswrapper[4758]: I1004 10:53:04.539512 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:05 crc kubenswrapper[4758]: I1004 10:53:05.539670 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:05 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:05 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:05 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:05 crc kubenswrapper[4758]: I1004 10:53:05.539736 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:05 crc kubenswrapper[4758]: I1004 10:53:05.557577 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-dwm2t container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Oct 04 10:53:05 crc kubenswrapper[4758]: I1004 10:53:05.557612 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dwm2t" podUID="dfcb1f56-7b71-4cf6-a92d-8f174194e697" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Oct 04 10:53:05 crc kubenswrapper[4758]: I1004 10:53:05.557658 4758 patch_prober.go:28] interesting pod/downloads-7954f5f757-dwm2t container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Oct 04 10:53:05 crc kubenswrapper[4758]: I1004 10:53:05.558125 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dwm2t" podUID="dfcb1f56-7b71-4cf6-a92d-8f174194e697" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Oct 04 10:53:06 crc kubenswrapper[4758]: I1004 10:53:06.538595 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:06 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:06 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:06 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:06 crc kubenswrapper[4758]: I1004 10:53:06.538690 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.168362 4758 patch_prober.go:28] interesting pod/console-f9d7485db-pqnhz container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.168713 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-pqnhz" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.539783 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:07 crc kubenswrapper[4758]: [-]has-synced failed: reason withheld Oct 04 10:53:07 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:07 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.539830 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.647860 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.658121 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e5adcca6-45b0-46ad-ac95-57cce5121dc9-metrics-certs\") pod \"network-metrics-daemon-srgqp\" (UID: \"e5adcca6-45b0-46ad-ac95-57cce5121dc9\") " pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:53:07 crc kubenswrapper[4758]: I1004 10:53:07.885638 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-srgqp" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.021127 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.155272 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b06830a3-62da-4416-9621-65fced1b4073-kube-api-access\") pod \"b06830a3-62da-4416-9621-65fced1b4073\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.155703 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b06830a3-62da-4416-9621-65fced1b4073-kubelet-dir\") pod \"b06830a3-62da-4416-9621-65fced1b4073\" (UID: \"b06830a3-62da-4416-9621-65fced1b4073\") " Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.156018 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b06830a3-62da-4416-9621-65fced1b4073-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b06830a3-62da-4416-9621-65fced1b4073" (UID: "b06830a3-62da-4416-9621-65fced1b4073"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.160290 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b06830a3-62da-4416-9621-65fced1b4073-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b06830a3-62da-4416-9621-65fced1b4073" (UID: "b06830a3-62da-4416-9621-65fced1b4073"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.256595 4758 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b06830a3-62da-4416-9621-65fced1b4073-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.256631 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b06830a3-62da-4416-9621-65fced1b4073-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.258945 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b06830a3-62da-4416-9621-65fced1b4073","Type":"ContainerDied","Data":"0f1566bd0e9b8a26958b00007106856053b6fb8c3fed5806e8945d4f10b149f1"} Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.258974 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f1566bd0e9b8a26958b00007106856053b6fb8c3fed5806e8945d4f10b149f1" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.259019 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.356025 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-srgqp"] Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.544249 4758 patch_prober.go:28] interesting pod/router-default-5444994796-hcsr7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:53:08 crc kubenswrapper[4758]: [+]has-synced ok Oct 04 10:53:08 crc kubenswrapper[4758]: [+]process-running ok Oct 04 10:53:08 crc kubenswrapper[4758]: healthz check failed Oct 04 10:53:08 crc kubenswrapper[4758]: I1004 10:53:08.544316 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hcsr7" podUID="818762b4-81b5-42fe-a8de-5908933306ad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:53:09 crc kubenswrapper[4758]: I1004 10:53:09.267097 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-srgqp" event={"ID":"e5adcca6-45b0-46ad-ac95-57cce5121dc9","Type":"ContainerStarted","Data":"87433285dea5c8b482e4ad53cbb8c3e0b27912b490b9ca7cefbe179a48403cab"} Oct 04 10:53:09 crc kubenswrapper[4758]: I1004 10:53:09.267732 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-srgqp" event={"ID":"e5adcca6-45b0-46ad-ac95-57cce5121dc9","Type":"ContainerStarted","Data":"05fd7f307ef723bfd80955d775d073740184b207a64e1d402097fb0afe761405"} Oct 04 10:53:09 crc kubenswrapper[4758]: I1004 10:53:09.540564 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:53:09 crc kubenswrapper[4758]: I1004 10:53:09.552515 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-hcsr7" Oct 04 10:53:15 crc kubenswrapper[4758]: I1004 10:53:15.569928 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dwm2t" Oct 04 10:53:17 crc kubenswrapper[4758]: I1004 10:53:17.172420 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:53:17 crc kubenswrapper[4758]: I1004 10:53:17.179424 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 10:53:17 crc kubenswrapper[4758]: I1004 10:53:17.342071 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:53:20 crc kubenswrapper[4758]: E1004 10:53:20.915026 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 10:53:20 crc kubenswrapper[4758]: E1004 10:53:20.915613 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vn5t6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6xncn_openshift-marketplace(447f9eb2-901e-42c0-813c-f03ac5f02916): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:53:20 crc kubenswrapper[4758]: E1004 10:53:20.916882 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6xncn" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" Oct 04 10:53:25 crc kubenswrapper[4758]: E1004 10:53:25.728829 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 10:53:25 crc kubenswrapper[4758]: E1004 10:53:25.729642 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5wcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-2pw5r_openshift-marketplace(8e82010e-ffbc-4e0e-b3aa-7e650d0e267e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:53:25 crc kubenswrapper[4758]: E1004 10:53:25.730953 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-2pw5r" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" Oct 04 10:53:26 crc kubenswrapper[4758]: I1004 10:53:26.854823 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-snbth" Oct 04 10:53:30 crc kubenswrapper[4758]: E1004 10:53:30.365325 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-2pw5r" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" Oct 04 10:53:30 crc kubenswrapper[4758]: E1004 10:53:30.365405 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6xncn" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" Oct 04 10:53:31 crc kubenswrapper[4758]: I1004 10:53:31.249798 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:53:31 crc kubenswrapper[4758]: I1004 10:53:31.250203 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:53:31 crc kubenswrapper[4758]: I1004 10:53:31.381517 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:53:31 crc kubenswrapper[4758]: E1004 10:53:31.698369 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 10:53:31 crc kubenswrapper[4758]: E1004 10:53:31.698515 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lprs9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-qgnlj_openshift-marketplace(688d59c7-04be-4f21-ae68-f134a6b72e66): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:53:31 crc kubenswrapper[4758]: E1004 10:53:31.701391 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-qgnlj" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" Oct 04 10:53:31 crc kubenswrapper[4758]: E1004 10:53:31.823822 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 10:53:31 crc kubenswrapper[4758]: E1004 10:53:31.824091 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-52744,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rnp22_openshift-marketplace(d3f8a403-d7ef-46d3-987e-f2d9b89f3172): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:53:31 crc kubenswrapper[4758]: E1004 10:53:31.825202 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rnp22" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" Oct 04 10:53:32 crc kubenswrapper[4758]: I1004 10:53:32.394357 4758 generic.go:334] "Generic (PLEG): container finished" podID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerID="fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0" exitCode=0 Oct 04 10:53:32 crc kubenswrapper[4758]: I1004 10:53:32.394427 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfvsd" event={"ID":"9c3a39ce-478e-4c93-b415-03700dcb90e1","Type":"ContainerDied","Data":"fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0"} Oct 04 10:53:32 crc kubenswrapper[4758]: I1004 10:53:32.400397 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-srgqp" event={"ID":"e5adcca6-45b0-46ad-ac95-57cce5121dc9","Type":"ContainerStarted","Data":"c378632c803df6f8b38cf04d322ada29178fc076cc6f58342d3da5d424bf5b77"} Oct 04 10:53:32 crc kubenswrapper[4758]: I1004 10:53:32.441858 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-srgqp" podStartSLOduration=167.44184008 podStartE2EDuration="2m47.44184008s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:53:32.43628264 +0000 UTC m=+189.728933529" watchObservedRunningTime="2025-10-04 10:53:32.44184008 +0000 UTC m=+189.734490979" Oct 04 10:53:34 crc kubenswrapper[4758]: E1004 10:53:34.454738 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-qgnlj" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" Oct 04 10:53:35 crc kubenswrapper[4758]: I1004 10:53:35.417945 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfvsd" event={"ID":"9c3a39ce-478e-4c93-b415-03700dcb90e1","Type":"ContainerStarted","Data":"5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118"} Oct 04 10:53:35 crc kubenswrapper[4758]: I1004 10:53:35.422483 4758 generic.go:334] "Generic (PLEG): container finished" podID="3bf4f130-658a-40c0-97c5-243d77950128" containerID="c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446" exitCode=0 Oct 04 10:53:35 crc kubenswrapper[4758]: I1004 10:53:35.422525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerDied","Data":"c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446"} Oct 04 10:53:35 crc kubenswrapper[4758]: I1004 10:53:35.435881 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bfvsd" podStartSLOduration=3.6620806569999997 podStartE2EDuration="42.435866011s" podCreationTimestamp="2025-10-04 10:52:53 +0000 UTC" firstStartedPulling="2025-10-04 10:52:55.931454889 +0000 UTC m=+153.224105778" lastFinishedPulling="2025-10-04 10:53:34.705240243 +0000 UTC m=+191.997891132" observedRunningTime="2025-10-04 10:53:35.434200939 +0000 UTC m=+192.726851838" watchObservedRunningTime="2025-10-04 10:53:35.435866011 +0000 UTC m=+192.728516900" Oct 04 10:53:40 crc kubenswrapper[4758]: I1004 10:53:40.455232 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerStarted","Data":"697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18"} Oct 04 10:53:40 crc kubenswrapper[4758]: I1004 10:53:40.460996 4758 generic.go:334] "Generic (PLEG): container finished" podID="45fc0b19-de79-477d-b2f4-db9786113587" containerID="55280c2bf537a65d7508327b9b6fa2676ef298ea49a6599b105a8c646db65ab7" exitCode=0 Oct 04 10:53:40 crc kubenswrapper[4758]: I1004 10:53:40.461067 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerDied","Data":"55280c2bf537a65d7508327b9b6fa2676ef298ea49a6599b105a8c646db65ab7"} Oct 04 10:53:40 crc kubenswrapper[4758]: I1004 10:53:40.466825 4758 generic.go:334] "Generic (PLEG): container finished" podID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerID="728a51e61c6079f6fe1a0a4a87b6024f3ff968117b43da605d62afb3d236959c" exitCode=0 Oct 04 10:53:40 crc kubenswrapper[4758]: I1004 10:53:40.466873 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xmk5z" event={"ID":"096babec-37b1-4321-a5f6-b3a614a0b49e","Type":"ContainerDied","Data":"728a51e61c6079f6fe1a0a4a87b6024f3ff968117b43da605d62afb3d236959c"} Oct 04 10:53:40 crc kubenswrapper[4758]: I1004 10:53:40.479647 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z5ck8" podStartSLOduration=3.7740700179999997 podStartE2EDuration="47.479627122s" podCreationTimestamp="2025-10-04 10:52:53 +0000 UTC" firstStartedPulling="2025-10-04 10:52:55.941608346 +0000 UTC m=+153.234259235" lastFinishedPulling="2025-10-04 10:53:39.64716543 +0000 UTC m=+196.939816339" observedRunningTime="2025-10-04 10:53:40.474679849 +0000 UTC m=+197.767330758" watchObservedRunningTime="2025-10-04 10:53:40.479627122 +0000 UTC m=+197.772278031" Oct 04 10:53:41 crc kubenswrapper[4758]: I1004 10:53:41.472631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xmk5z" event={"ID":"096babec-37b1-4321-a5f6-b3a614a0b49e","Type":"ContainerStarted","Data":"d6a20bc6d05c4d90f67d720e3f3f8b8e588aefad37402e67a59eebbd3dd90583"} Oct 04 10:53:41 crc kubenswrapper[4758]: I1004 10:53:41.475445 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerStarted","Data":"4f08d122467ac9bb5accce2ca4f04c1275835023492ba2c18de6443ee86965f1"} Oct 04 10:53:41 crc kubenswrapper[4758]: I1004 10:53:41.492485 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xmk5z" podStartSLOduration=4.69426454 podStartE2EDuration="46.492466985s" podCreationTimestamp="2025-10-04 10:52:55 +0000 UTC" firstStartedPulling="2025-10-04 10:52:59.186126032 +0000 UTC m=+156.478776921" lastFinishedPulling="2025-10-04 10:53:40.984328477 +0000 UTC m=+198.276979366" observedRunningTime="2025-10-04 10:53:41.490616895 +0000 UTC m=+198.783267784" watchObservedRunningTime="2025-10-04 10:53:41.492466985 +0000 UTC m=+198.785117874" Oct 04 10:53:41 crc kubenswrapper[4758]: I1004 10:53:41.513104 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c8zz9" podStartSLOduration=4.780405968 podStartE2EDuration="46.513083923s" podCreationTimestamp="2025-10-04 10:52:55 +0000 UTC" firstStartedPulling="2025-10-04 10:52:59.18285764 +0000 UTC m=+156.475508529" lastFinishedPulling="2025-10-04 10:53:40.915535595 +0000 UTC m=+198.208186484" observedRunningTime="2025-10-04 10:53:41.50903225 +0000 UTC m=+198.801683129" watchObservedRunningTime="2025-10-04 10:53:41.513083923 +0000 UTC m=+198.805734812" Oct 04 10:53:43 crc kubenswrapper[4758]: I1004 10:53:43.555391 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:53:43 crc kubenswrapper[4758]: I1004 10:53:43.556615 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:53:43 crc kubenswrapper[4758]: I1004 10:53:43.854562 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:53:43 crc kubenswrapper[4758]: I1004 10:53:43.982335 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:53:43 crc kubenswrapper[4758]: I1004 10:53:43.982385 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:53:44 crc kubenswrapper[4758]: I1004 10:53:44.042322 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:53:44 crc kubenswrapper[4758]: I1004 10:53:44.564836 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:53:44 crc kubenswrapper[4758]: I1004 10:53:44.573358 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:53:45 crc kubenswrapper[4758]: I1004 10:53:45.515933 4758 generic.go:334] "Generic (PLEG): container finished" podID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerID="f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628" exitCode=0 Oct 04 10:53:45 crc kubenswrapper[4758]: I1004 10:53:45.516003 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pw5r" event={"ID":"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e","Type":"ContainerDied","Data":"f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628"} Oct 04 10:53:45 crc kubenswrapper[4758]: I1004 10:53:45.719287 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:53:45 crc kubenswrapper[4758]: I1004 10:53:45.719626 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:53:45 crc kubenswrapper[4758]: I1004 10:53:45.773317 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.137949 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.138035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.184912 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.288785 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bfvsd"] Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.524719 4758 generic.go:334] "Generic (PLEG): container finished" podID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerID="85e03319a32408086040acfa1d5949cce0edef585cf35f47d9349077c7a2947d" exitCode=0 Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.524791 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xncn" event={"ID":"447f9eb2-901e-42c0-813c-f03ac5f02916","Type":"ContainerDied","Data":"85e03319a32408086040acfa1d5949cce0edef585cf35f47d9349077c7a2947d"} Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.527077 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pw5r" event={"ID":"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e","Type":"ContainerStarted","Data":"a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946"} Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.527500 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bfvsd" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="registry-server" containerID="cri-o://5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118" gracePeriod=2 Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.576494 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2pw5r" podStartSLOduration=3.517905798 podStartE2EDuration="53.576476061s" podCreationTimestamp="2025-10-04 10:52:53 +0000 UTC" firstStartedPulling="2025-10-04 10:52:55.988368956 +0000 UTC m=+153.281019845" lastFinishedPulling="2025-10-04 10:53:46.046939219 +0000 UTC m=+203.339590108" observedRunningTime="2025-10-04 10:53:46.572749849 +0000 UTC m=+203.865400738" watchObservedRunningTime="2025-10-04 10:53:46.576476061 +0000 UTC m=+203.869126950" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.586291 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.586723 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.946133 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.973257 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-catalog-content\") pod \"9c3a39ce-478e-4c93-b415-03700dcb90e1\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.973307 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2zbx\" (UniqueName: \"kubernetes.io/projected/9c3a39ce-478e-4c93-b415-03700dcb90e1-kube-api-access-f2zbx\") pod \"9c3a39ce-478e-4c93-b415-03700dcb90e1\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.973351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-utilities\") pod \"9c3a39ce-478e-4c93-b415-03700dcb90e1\" (UID: \"9c3a39ce-478e-4c93-b415-03700dcb90e1\") " Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.975700 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-utilities" (OuterVolumeSpecName: "utilities") pod "9c3a39ce-478e-4c93-b415-03700dcb90e1" (UID: "9c3a39ce-478e-4c93-b415-03700dcb90e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:46 crc kubenswrapper[4758]: I1004 10:53:46.979189 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c3a39ce-478e-4c93-b415-03700dcb90e1-kube-api-access-f2zbx" (OuterVolumeSpecName: "kube-api-access-f2zbx") pod "9c3a39ce-478e-4c93-b415-03700dcb90e1" (UID: "9c3a39ce-478e-4c93-b415-03700dcb90e1"). InnerVolumeSpecName "kube-api-access-f2zbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.024637 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c3a39ce-478e-4c93-b415-03700dcb90e1" (UID: "9c3a39ce-478e-4c93-b415-03700dcb90e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.074188 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.074222 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2zbx\" (UniqueName: \"kubernetes.io/projected/9c3a39ce-478e-4c93-b415-03700dcb90e1-kube-api-access-f2zbx\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.074232 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c3a39ce-478e-4c93-b415-03700dcb90e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.537773 4758 generic.go:334] "Generic (PLEG): container finished" podID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerID="5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118" exitCode=0 Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.537829 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfvsd" event={"ID":"9c3a39ce-478e-4c93-b415-03700dcb90e1","Type":"ContainerDied","Data":"5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118"} Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.537896 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bfvsd" event={"ID":"9c3a39ce-478e-4c93-b415-03700dcb90e1","Type":"ContainerDied","Data":"78d66a881117e1dc76677e429ec488379ca3a22cede5ce15582ae34b05d16895"} Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.537922 4758 scope.go:117] "RemoveContainer" containerID="5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.537999 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bfvsd" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.540332 4758 generic.go:334] "Generic (PLEG): container finished" podID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerID="f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04" exitCode=0 Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.540431 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnp22" event={"ID":"d3f8a403-d7ef-46d3-987e-f2d9b89f3172","Type":"ContainerDied","Data":"f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04"} Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.562887 4758 scope.go:117] "RemoveContainer" containerID="fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.592729 4758 scope.go:117] "RemoveContainer" containerID="dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.592742 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bfvsd"] Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.595160 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bfvsd"] Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.613036 4758 scope.go:117] "RemoveContainer" containerID="5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118" Oct 04 10:53:47 crc kubenswrapper[4758]: E1004 10:53:47.614997 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118\": container with ID starting with 5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118 not found: ID does not exist" containerID="5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.615038 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118"} err="failed to get container status \"5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118\": rpc error: code = NotFound desc = could not find container \"5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118\": container with ID starting with 5e630db2805f262316036e577aca7d1d381159686e49efe2ab3f5db062a4a118 not found: ID does not exist" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.615088 4758 scope.go:117] "RemoveContainer" containerID="fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0" Oct 04 10:53:47 crc kubenswrapper[4758]: E1004 10:53:47.616679 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0\": container with ID starting with fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0 not found: ID does not exist" containerID="fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.616761 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0"} err="failed to get container status \"fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0\": rpc error: code = NotFound desc = could not find container \"fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0\": container with ID starting with fc10cd631940b846ec284fa09ede4f489cc45f25ac1a88fc181d4adbcdf584a0 not found: ID does not exist" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.616823 4758 scope.go:117] "RemoveContainer" containerID="dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e" Oct 04 10:53:47 crc kubenswrapper[4758]: E1004 10:53:47.617316 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e\": container with ID starting with dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e not found: ID does not exist" containerID="dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e" Oct 04 10:53:47 crc kubenswrapper[4758]: I1004 10:53:47.617349 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e"} err="failed to get container status \"dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e\": rpc error: code = NotFound desc = could not find container \"dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e\": container with ID starting with dbdddd9e4b3895e76de910c73a96c6ecd4ff81c2e3567645d6100df77892c67e not found: ID does not exist" Oct 04 10:53:48 crc kubenswrapper[4758]: I1004 10:53:48.548417 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xncn" event={"ID":"447f9eb2-901e-42c0-813c-f03ac5f02916","Type":"ContainerStarted","Data":"fa78650332def9d0743ce974624f3669282a1e8b6e290ff6d29d6505c9a4a317"} Oct 04 10:53:49 crc kubenswrapper[4758]: I1004 10:53:49.291799 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8zz9"] Oct 04 10:53:49 crc kubenswrapper[4758]: I1004 10:53:49.292175 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c8zz9" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="registry-server" containerID="cri-o://4f08d122467ac9bb5accce2ca4f04c1275835023492ba2c18de6443ee86965f1" gracePeriod=2 Oct 04 10:53:49 crc kubenswrapper[4758]: I1004 10:53:49.345769 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" path="/var/lib/kubelet/pods/9c3a39ce-478e-4c93-b415-03700dcb90e1/volumes" Oct 04 10:53:51 crc kubenswrapper[4758]: I1004 10:53:51.568820 4758 generic.go:334] "Generic (PLEG): container finished" podID="45fc0b19-de79-477d-b2f4-db9786113587" containerID="4f08d122467ac9bb5accce2ca4f04c1275835023492ba2c18de6443ee86965f1" exitCode=0 Oct 04 10:53:51 crc kubenswrapper[4758]: I1004 10:53:51.568889 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerDied","Data":"4f08d122467ac9bb5accce2ca4f04c1275835023492ba2c18de6443ee86965f1"} Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.638431 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.656937 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-catalog-content\") pod \"45fc0b19-de79-477d-b2f4-db9786113587\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.657057 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzmtj\" (UniqueName: \"kubernetes.io/projected/45fc0b19-de79-477d-b2f4-db9786113587-kube-api-access-kzmtj\") pod \"45fc0b19-de79-477d-b2f4-db9786113587\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.657135 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-utilities\") pod \"45fc0b19-de79-477d-b2f4-db9786113587\" (UID: \"45fc0b19-de79-477d-b2f4-db9786113587\") " Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.664577 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-utilities" (OuterVolumeSpecName: "utilities") pod "45fc0b19-de79-477d-b2f4-db9786113587" (UID: "45fc0b19-de79-477d-b2f4-db9786113587"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.673343 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45fc0b19-de79-477d-b2f4-db9786113587-kube-api-access-kzmtj" (OuterVolumeSpecName: "kube-api-access-kzmtj") pod "45fc0b19-de79-477d-b2f4-db9786113587" (UID: "45fc0b19-de79-477d-b2f4-db9786113587"). InnerVolumeSpecName "kube-api-access-kzmtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.678611 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6xncn" podStartSLOduration=9.308735055 podStartE2EDuration="1m0.678593174s" podCreationTimestamp="2025-10-04 10:52:53 +0000 UTC" firstStartedPulling="2025-10-04 10:52:55.958123713 +0000 UTC m=+153.250774602" lastFinishedPulling="2025-10-04 10:53:47.327981822 +0000 UTC m=+204.620632721" observedRunningTime="2025-10-04 10:53:49.582178771 +0000 UTC m=+206.874829710" watchObservedRunningTime="2025-10-04 10:53:53.678593174 +0000 UTC m=+210.971244073" Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.758906 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:53 crc kubenswrapper[4758]: I1004 10:53:53.758938 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzmtj\" (UniqueName: \"kubernetes.io/projected/45fc0b19-de79-477d-b2f4-db9786113587-kube-api-access-kzmtj\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.026899 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.026952 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.066614 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.183972 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.184285 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.225744 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.504989 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45fc0b19-de79-477d-b2f4-db9786113587" (UID: "45fc0b19-de79-477d-b2f4-db9786113587"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.569510 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45fc0b19-de79-477d-b2f4-db9786113587-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.589024 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c8zz9" event={"ID":"45fc0b19-de79-477d-b2f4-db9786113587","Type":"ContainerDied","Data":"e9ce8a9562f20f7a18dfbec0bc8dd6acbdedda292ba277e70feb14c6d0010564"} Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.589048 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c8zz9" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.589090 4758 scope.go:117] "RemoveContainer" containerID="4f08d122467ac9bb5accce2ca4f04c1275835023492ba2c18de6443ee86965f1" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.591323 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnp22" event={"ID":"d3f8a403-d7ef-46d3-987e-f2d9b89f3172","Type":"ContainerStarted","Data":"f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2"} Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.606649 4758 scope.go:117] "RemoveContainer" containerID="55280c2bf537a65d7508327b9b6fa2676ef298ea49a6599b105a8c646db65ab7" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.620436 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8zz9"] Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.628053 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c8zz9"] Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.643026 4758 scope.go:117] "RemoveContainer" containerID="87129a6c82e52a0acef61af70d9c34a9f3bc8d34eb3f3d9cf9a7ee98b4da7dfa" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.643873 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:53:54 crc kubenswrapper[4758]: I1004 10:53:54.673454 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:53:55 crc kubenswrapper[4758]: I1004 10:53:55.335777 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45fc0b19-de79-477d-b2f4-db9786113587" path="/var/lib/kubelet/pods/45fc0b19-de79-477d-b2f4-db9786113587/volumes" Oct 04 10:53:55 crc kubenswrapper[4758]: I1004 10:53:55.598736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerStarted","Data":"2f152fd4d185fe0dc19780c7cf55d13f26c412aede9a424fae1b71c65e9839ad"} Oct 04 10:53:55 crc kubenswrapper[4758]: I1004 10:53:55.641864 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rnp22" podStartSLOduration=7.067047575 podStartE2EDuration="59.641838708s" podCreationTimestamp="2025-10-04 10:52:56 +0000 UTC" firstStartedPulling="2025-10-04 10:53:00.197053418 +0000 UTC m=+157.489704307" lastFinishedPulling="2025-10-04 10:53:52.771844511 +0000 UTC m=+210.064495440" observedRunningTime="2025-10-04 10:53:55.635427268 +0000 UTC m=+212.928078167" watchObservedRunningTime="2025-10-04 10:53:55.641838708 +0000 UTC m=+212.934489637" Oct 04 10:53:56 crc kubenswrapper[4758]: I1004 10:53:56.606530 4758 generic.go:334] "Generic (PLEG): container finished" podID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerID="2f152fd4d185fe0dc19780c7cf55d13f26c412aede9a424fae1b71c65e9839ad" exitCode=0 Oct 04 10:53:56 crc kubenswrapper[4758]: I1004 10:53:56.607142 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerDied","Data":"2f152fd4d185fe0dc19780c7cf55d13f26c412aede9a424fae1b71c65e9839ad"} Oct 04 10:53:57 crc kubenswrapper[4758]: I1004 10:53:57.775431 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:53:57 crc kubenswrapper[4758]: I1004 10:53:57.775964 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.085852 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2pw5r"] Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.086043 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2pw5r" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="registry-server" containerID="cri-o://a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946" gracePeriod=2 Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.574668 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.618163 4758 generic.go:334] "Generic (PLEG): container finished" podID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerID="a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946" exitCode=0 Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.618206 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2pw5r" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.618261 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pw5r" event={"ID":"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e","Type":"ContainerDied","Data":"a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946"} Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.618309 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2pw5r" event={"ID":"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e","Type":"ContainerDied","Data":"37da6e968180506a82ad20d3eaa7310e1e4fc23e352b48cbb34fd0c8491998bf"} Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.618328 4758 scope.go:117] "RemoveContainer" containerID="a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.622531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerStarted","Data":"8b1057789c51acaa9dc59c7af3bdc131bee8f3136d77ca30bb8610340fd5ac06"} Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.632722 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-utilities\") pod \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.632859 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5wcc\" (UniqueName: \"kubernetes.io/projected/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-kube-api-access-b5wcc\") pod \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.632896 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-catalog-content\") pod \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\" (UID: \"8e82010e-ffbc-4e0e-b3aa-7e650d0e267e\") " Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.633610 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-utilities" (OuterVolumeSpecName: "utilities") pod "8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" (UID: "8e82010e-ffbc-4e0e-b3aa-7e650d0e267e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.638443 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qgnlj" podStartSLOduration=5.355805722 podStartE2EDuration="1m2.638429479s" podCreationTimestamp="2025-10-04 10:52:56 +0000 UTC" firstStartedPulling="2025-10-04 10:53:00.196655088 +0000 UTC m=+157.489305977" lastFinishedPulling="2025-10-04 10:53:57.479278815 +0000 UTC m=+214.771929734" observedRunningTime="2025-10-04 10:53:58.636003189 +0000 UTC m=+215.928654078" watchObservedRunningTime="2025-10-04 10:53:58.638429479 +0000 UTC m=+215.931080368" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.640895 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-kube-api-access-b5wcc" (OuterVolumeSpecName: "kube-api-access-b5wcc") pod "8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" (UID: "8e82010e-ffbc-4e0e-b3aa-7e650d0e267e"). InnerVolumeSpecName "kube-api-access-b5wcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.645063 4758 scope.go:117] "RemoveContainer" containerID="f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.662387 4758 scope.go:117] "RemoveContainer" containerID="abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.688226 4758 scope.go:117] "RemoveContainer" containerID="a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946" Oct 04 10:53:58 crc kubenswrapper[4758]: E1004 10:53:58.690430 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946\": container with ID starting with a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946 not found: ID does not exist" containerID="a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.690457 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946"} err="failed to get container status \"a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946\": rpc error: code = NotFound desc = could not find container \"a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946\": container with ID starting with a9e0b0037b6eae5f39a8b741a16fd418511a510532343c366df70d83d73af946 not found: ID does not exist" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.690476 4758 scope.go:117] "RemoveContainer" containerID="f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628" Oct 04 10:53:58 crc kubenswrapper[4758]: E1004 10:53:58.692393 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628\": container with ID starting with f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628 not found: ID does not exist" containerID="f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.692411 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628"} err="failed to get container status \"f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628\": rpc error: code = NotFound desc = could not find container \"f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628\": container with ID starting with f90801aedfa36d077ba079772e66e8e160ba07385d7f1ce905ad29d79182b628 not found: ID does not exist" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.692426 4758 scope.go:117] "RemoveContainer" containerID="abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96" Oct 04 10:53:58 crc kubenswrapper[4758]: E1004 10:53:58.692774 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96\": container with ID starting with abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96 not found: ID does not exist" containerID="abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.692790 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96"} err="failed to get container status \"abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96\": rpc error: code = NotFound desc = could not find container \"abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96\": container with ID starting with abcaa6359491edefabcf41c8c971842c7322a581ae718d6a75dc76d3100e3b96 not found: ID does not exist" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.698204 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" (UID: "8e82010e-ffbc-4e0e-b3aa-7e650d0e267e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.734356 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5wcc\" (UniqueName: \"kubernetes.io/projected/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-kube-api-access-b5wcc\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.734381 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.734391 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.827309 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rnp22" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="registry-server" probeResult="failure" output=< Oct 04 10:53:58 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 10:53:58 crc kubenswrapper[4758]: > Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.946016 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2pw5r"] Oct 04 10:53:58 crc kubenswrapper[4758]: I1004 10:53:58.954319 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2pw5r"] Oct 04 10:53:59 crc kubenswrapper[4758]: I1004 10:53:59.333878 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" path="/var/lib/kubelet/pods/8e82010e-ffbc-4e0e-b3aa-7e650d0e267e/volumes" Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.249810 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.250227 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.250294 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.251131 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.251228 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96" gracePeriod=600 Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.644089 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96" exitCode=0 Oct 04 10:54:01 crc kubenswrapper[4758]: I1004 10:54:01.644195 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96"} Oct 04 10:54:03 crc kubenswrapper[4758]: I1004 10:54:03.658112 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"770a65d2316af3d496832d48f9bf7d288866132c38f07143143f0ef4b5eb238f"} Oct 04 10:54:07 crc kubenswrapper[4758]: I1004 10:54:07.761523 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:54:07 crc kubenswrapper[4758]: I1004 10:54:07.762845 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:54:07 crc kubenswrapper[4758]: I1004 10:54:07.814450 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:54:07 crc kubenswrapper[4758]: I1004 10:54:07.840335 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:54:07 crc kubenswrapper[4758]: I1004 10:54:07.876712 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:54:08 crc kubenswrapper[4758]: I1004 10:54:08.725153 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:54:09 crc kubenswrapper[4758]: I1004 10:54:09.044773 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnp22"] Oct 04 10:54:09 crc kubenswrapper[4758]: I1004 10:54:09.685870 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rnp22" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="registry-server" containerID="cri-o://f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2" gracePeriod=2 Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.184576 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.306004 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52744\" (UniqueName: \"kubernetes.io/projected/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-kube-api-access-52744\") pod \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.306062 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-utilities\") pod \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.306128 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-catalog-content\") pod \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\" (UID: \"d3f8a403-d7ef-46d3-987e-f2d9b89f3172\") " Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.307021 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-utilities" (OuterVolumeSpecName: "utilities") pod "d3f8a403-d7ef-46d3-987e-f2d9b89f3172" (UID: "d3f8a403-d7ef-46d3-987e-f2d9b89f3172"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.314038 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-kube-api-access-52744" (OuterVolumeSpecName: "kube-api-access-52744") pod "d3f8a403-d7ef-46d3-987e-f2d9b89f3172" (UID: "d3f8a403-d7ef-46d3-987e-f2d9b89f3172"). InnerVolumeSpecName "kube-api-access-52744". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.390244 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3f8a403-d7ef-46d3-987e-f2d9b89f3172" (UID: "d3f8a403-d7ef-46d3-987e-f2d9b89f3172"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.407661 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.407726 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52744\" (UniqueName: \"kubernetes.io/projected/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-kube-api-access-52744\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.407744 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3f8a403-d7ef-46d3-987e-f2d9b89f3172-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.694975 4758 generic.go:334] "Generic (PLEG): container finished" podID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerID="f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2" exitCode=0 Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.696437 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnp22" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.704800 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnp22" event={"ID":"d3f8a403-d7ef-46d3-987e-f2d9b89f3172","Type":"ContainerDied","Data":"f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2"} Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.704874 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnp22" event={"ID":"d3f8a403-d7ef-46d3-987e-f2d9b89f3172","Type":"ContainerDied","Data":"6db9d2767070633c3f3793f845fa360a92daec9e79d34b764a277f7d6af33d1f"} Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.704895 4758 scope.go:117] "RemoveContainer" containerID="f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.721112 4758 scope.go:117] "RemoveContainer" containerID="f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.741961 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnp22"] Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.749538 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rnp22"] Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.755892 4758 scope.go:117] "RemoveContainer" containerID="a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.769432 4758 scope.go:117] "RemoveContainer" containerID="f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2" Oct 04 10:54:10 crc kubenswrapper[4758]: E1004 10:54:10.770205 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2\": container with ID starting with f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2 not found: ID does not exist" containerID="f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.770481 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2"} err="failed to get container status \"f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2\": rpc error: code = NotFound desc = could not find container \"f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2\": container with ID starting with f435a21265b72093cc47d07c54d773619a5ea966a639dde88eaa6931e4d871e2 not found: ID does not exist" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.770676 4758 scope.go:117] "RemoveContainer" containerID="f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04" Oct 04 10:54:10 crc kubenswrapper[4758]: E1004 10:54:10.771399 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04\": container with ID starting with f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04 not found: ID does not exist" containerID="f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.771442 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04"} err="failed to get container status \"f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04\": rpc error: code = NotFound desc = could not find container \"f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04\": container with ID starting with f59ca543829d4277c701b9d3b5cf5ad17795934c2111e4007e607b28803e3b04 not found: ID does not exist" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.771473 4758 scope.go:117] "RemoveContainer" containerID="a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2" Oct 04 10:54:10 crc kubenswrapper[4758]: E1004 10:54:10.771948 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2\": container with ID starting with a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2 not found: ID does not exist" containerID="a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2" Oct 04 10:54:10 crc kubenswrapper[4758]: I1004 10:54:10.772209 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2"} err="failed to get container status \"a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2\": rpc error: code = NotFound desc = could not find container \"a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2\": container with ID starting with a6c96bd951e22ac039460583c873d8745971f33663fc26f1cd7a091c05ce47f2 not found: ID does not exist" Oct 04 10:54:11 crc kubenswrapper[4758]: I1004 10:54:11.331809 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" path="/var/lib/kubelet/pods/d3f8a403-d7ef-46d3-987e-f2d9b89f3172/volumes" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.070223 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4k6bj"] Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.070937 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.070952 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.070964 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.070972 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.070984 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06830a3-62da-4416-9621-65fced1b4073" containerName="pruner" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.070992 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06830a3-62da-4416-9621-65fced1b4073" containerName="pruner" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071008 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071016 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071032 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071041 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071053 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071061 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071075 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071083 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071095 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071123 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071134 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071143 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071153 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071161 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071174 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071182 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="extract-content" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071191 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071198 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="extract-utilities" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071209 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071217 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: E1004 10:54:20.071230 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50" containerName="pruner" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071238 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50" containerName="pruner" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071345 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="45fc0b19-de79-477d-b2f4-db9786113587" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071363 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b676bd0c-3f4f-434d-bdfe-24f3f5c0ef50" containerName="pruner" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071372 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e82010e-ffbc-4e0e-b3aa-7e650d0e267e" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071384 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3a39ce-478e-4c93-b415-03700dcb90e1" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071395 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f8a403-d7ef-46d3-987e-f2d9b89f3172" containerName="registry-server" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071407 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b06830a3-62da-4416-9621-65fced1b4073" containerName="pruner" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.071821 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.088790 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4k6bj"] Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.127333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0935d70c-d579-4663-ab53-fd152ecc9e1b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.127566 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-registry-tls\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.127661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0935d70c-d579-4663-ab53-fd152ecc9e1b-trusted-ca\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.127751 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0935d70c-d579-4663-ab53-fd152ecc9e1b-registry-certificates\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.127834 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-bound-sa-token\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.127917 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0935d70c-d579-4663-ab53-fd152ecc9e1b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.128005 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crxt\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-kube-api-access-2crxt\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.128194 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.166664 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.228967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0935d70c-d579-4663-ab53-fd152ecc9e1b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229022 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crxt\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-kube-api-access-2crxt\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229050 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0935d70c-d579-4663-ab53-fd152ecc9e1b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229080 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-registry-tls\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229119 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0935d70c-d579-4663-ab53-fd152ecc9e1b-trusted-ca\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229142 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0935d70c-d579-4663-ab53-fd152ecc9e1b-registry-certificates\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229163 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-bound-sa-token\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.229582 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0935d70c-d579-4663-ab53-fd152ecc9e1b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.230641 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0935d70c-d579-4663-ab53-fd152ecc9e1b-trusted-ca\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.230788 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0935d70c-d579-4663-ab53-fd152ecc9e1b-registry-certificates\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.235008 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-registry-tls\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.237210 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0935d70c-d579-4663-ab53-fd152ecc9e1b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.243823 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-bound-sa-token\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.249664 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crxt\" (UniqueName: \"kubernetes.io/projected/0935d70c-d579-4663-ab53-fd152ecc9e1b-kube-api-access-2crxt\") pod \"image-registry-66df7c8f76-4k6bj\" (UID: \"0935d70c-d579-4663-ab53-fd152ecc9e1b\") " pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.389160 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:20 crc kubenswrapper[4758]: I1004 10:54:20.794555 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4k6bj"] Oct 04 10:54:20 crc kubenswrapper[4758]: W1004 10:54:20.803655 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0935d70c_d579_4663_ab53_fd152ecc9e1b.slice/crio-ace331772eee2b0d8d8b7931dfdab2979bef0edd588194851a0a835447522c56 WatchSource:0}: Error finding container ace331772eee2b0d8d8b7931dfdab2979bef0edd588194851a0a835447522c56: Status 404 returned error can't find the container with id ace331772eee2b0d8d8b7931dfdab2979bef0edd588194851a0a835447522c56 Oct 04 10:54:21 crc kubenswrapper[4758]: I1004 10:54:21.757112 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" event={"ID":"0935d70c-d579-4663-ab53-fd152ecc9e1b","Type":"ContainerStarted","Data":"5e0d41bd5dd9938fa2625d584c886d45713bf685326c27ff9305e89e92027301"} Oct 04 10:54:21 crc kubenswrapper[4758]: I1004 10:54:21.757419 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" event={"ID":"0935d70c-d579-4663-ab53-fd152ecc9e1b","Type":"ContainerStarted","Data":"ace331772eee2b0d8d8b7931dfdab2979bef0edd588194851a0a835447522c56"} Oct 04 10:54:21 crc kubenswrapper[4758]: I1004 10:54:21.757538 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:21 crc kubenswrapper[4758]: I1004 10:54:21.783692 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" podStartSLOduration=1.783670194 podStartE2EDuration="1.783670194s" podCreationTimestamp="2025-10-04 10:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:21.779686413 +0000 UTC m=+239.072337332" watchObservedRunningTime="2025-10-04 10:54:21.783670194 +0000 UTC m=+239.076321123" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.546886 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z5ck8"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.549441 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z5ck8" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="registry-server" containerID="cri-o://697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18" gracePeriod=30 Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.552779 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6xncn"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.553333 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6xncn" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="registry-server" containerID="cri-o://fa78650332def9d0743ce974624f3669282a1e8b6e290ff6d29d6505c9a4a317" gracePeriod=30 Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.570350 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s4v4x"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.570888 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" containerID="cri-o://0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21" gracePeriod=30 Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.589620 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qhg7v"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.590467 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.592687 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xmk5z"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.593050 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xmk5z" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="registry-server" containerID="cri-o://d6a20bc6d05c4d90f67d720e3f3f8b8e588aefad37402e67a59eebbd3dd90583" gracePeriod=30 Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.609578 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qhg7v"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.612413 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qgnlj"] Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.612683 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qgnlj" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="registry-server" containerID="cri-o://8b1057789c51acaa9dc59c7af3bdc131bee8f3136d77ca30bb8610340fd5ac06" gracePeriod=30 Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.675541 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96f085f8-d3c4-4e51-ab9b-ac854a355e58-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.675589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96f085f8-d3c4-4e51-ab9b-ac854a355e58-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.675630 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkfvv\" (UniqueName: \"kubernetes.io/projected/96f085f8-d3c4-4e51-ab9b-ac854a355e58-kube-api-access-wkfvv\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.776642 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96f085f8-d3c4-4e51-ab9b-ac854a355e58-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.776684 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96f085f8-d3c4-4e51-ab9b-ac854a355e58-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.776756 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkfvv\" (UniqueName: \"kubernetes.io/projected/96f085f8-d3c4-4e51-ab9b-ac854a355e58-kube-api-access-wkfvv\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.778145 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96f085f8-d3c4-4e51-ab9b-ac854a355e58-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.783806 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96f085f8-d3c4-4e51-ab9b-ac854a355e58-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.791561 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkfvv\" (UniqueName: \"kubernetes.io/projected/96f085f8-d3c4-4e51-ab9b-ac854a355e58-kube-api-access-wkfvv\") pod \"marketplace-operator-79b997595-qhg7v\" (UID: \"96f085f8-d3c4-4e51-ab9b-ac854a355e58\") " pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:29 crc kubenswrapper[4758]: I1004 10:54:29.915970 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.087920 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8zmq8"] Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.507732 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qhg7v"] Oct 04 10:54:30 crc kubenswrapper[4758]: W1004 10:54:30.520934 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96f085f8_d3c4_4e51_ab9b_ac854a355e58.slice/crio-747b7f4234fb0a4565c20660ab2d9b092c7c069b45d6c84dead7562e5f37dc83 WatchSource:0}: Error finding container 747b7f4234fb0a4565c20660ab2d9b092c7c069b45d6c84dead7562e5f37dc83: Status 404 returned error can't find the container with id 747b7f4234fb0a4565c20660ab2d9b092c7c069b45d6c84dead7562e5f37dc83 Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.632404 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.693718 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-operator-metrics\") pod \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.693808 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-trusted-ca\") pod \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.693842 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mbtj\" (UniqueName: \"kubernetes.io/projected/583b0f46-ae37-4ba6-9d33-81fd3b22f633-kube-api-access-5mbtj\") pod \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\" (UID: \"583b0f46-ae37-4ba6-9d33-81fd3b22f633\") " Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.694853 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "583b0f46-ae37-4ba6-9d33-81fd3b22f633" (UID: "583b0f46-ae37-4ba6-9d33-81fd3b22f633"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.702003 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "583b0f46-ae37-4ba6-9d33-81fd3b22f633" (UID: "583b0f46-ae37-4ba6-9d33-81fd3b22f633"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.719943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583b0f46-ae37-4ba6-9d33-81fd3b22f633-kube-api-access-5mbtj" (OuterVolumeSpecName: "kube-api-access-5mbtj") pod "583b0f46-ae37-4ba6-9d33-81fd3b22f633" (UID: "583b0f46-ae37-4ba6-9d33-81fd3b22f633"). InnerVolumeSpecName "kube-api-access-5mbtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.796289 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.796315 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mbtj\" (UniqueName: \"kubernetes.io/projected/583b0f46-ae37-4ba6-9d33-81fd3b22f633-kube-api-access-5mbtj\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.796327 4758 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/583b0f46-ae37-4ba6-9d33-81fd3b22f633-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.819227 4758 generic.go:334] "Generic (PLEG): container finished" podID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerID="fa78650332def9d0743ce974624f3669282a1e8b6e290ff6d29d6505c9a4a317" exitCode=0 Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.819287 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xncn" event={"ID":"447f9eb2-901e-42c0-813c-f03ac5f02916","Type":"ContainerDied","Data":"fa78650332def9d0743ce974624f3669282a1e8b6e290ff6d29d6505c9a4a317"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.824525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" event={"ID":"96f085f8-d3c4-4e51-ab9b-ac854a355e58","Type":"ContainerStarted","Data":"b1b9c684832c7e7a456f992e175e9e883f0a34e78986182b4cf270235ec854dd"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.824569 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" event={"ID":"96f085f8-d3c4-4e51-ab9b-ac854a355e58","Type":"ContainerStarted","Data":"747b7f4234fb0a4565c20660ab2d9b092c7c069b45d6c84dead7562e5f37dc83"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.824585 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.837711 4758 generic.go:334] "Generic (PLEG): container finished" podID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerID="d6a20bc6d05c4d90f67d720e3f3f8b8e588aefad37402e67a59eebbd3dd90583" exitCode=0 Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.837828 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xmk5z" event={"ID":"096babec-37b1-4321-a5f6-b3a614a0b49e","Type":"ContainerDied","Data":"d6a20bc6d05c4d90f67d720e3f3f8b8e588aefad37402e67a59eebbd3dd90583"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.838287 4758 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qhg7v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.838327 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" podUID="96f085f8-d3c4-4e51-ab9b-ac854a355e58" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.842247 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.847881 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" podStartSLOduration=1.847859944 podStartE2EDuration="1.847859944s" podCreationTimestamp="2025-10-04 10:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:30.841494494 +0000 UTC m=+248.134145383" watchObservedRunningTime="2025-10-04 10:54:30.847859944 +0000 UTC m=+248.140510833" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.848410 4758 generic.go:334] "Generic (PLEG): container finished" podID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerID="8b1057789c51acaa9dc59c7af3bdc131bee8f3136d77ca30bb8610340fd5ac06" exitCode=0 Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.848507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerDied","Data":"8b1057789c51acaa9dc59c7af3bdc131bee8f3136d77ca30bb8610340fd5ac06"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.880437 4758 generic.go:334] "Generic (PLEG): container finished" podID="3bf4f130-658a-40c0-97c5-243d77950128" containerID="697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18" exitCode=0 Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.880691 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerDied","Data":"697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.880774 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5ck8" event={"ID":"3bf4f130-658a-40c0-97c5-243d77950128","Type":"ContainerDied","Data":"87c5ef0ad85aa3f4757bb167dfedde0c9a90d0453433737cde2ecf42f3358f0a"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.880881 4758 scope.go:117] "RemoveContainer" containerID="697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.881071 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5ck8" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.885989 4758 generic.go:334] "Generic (PLEG): container finished" podID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerID="0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21" exitCode=0 Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.886122 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" event={"ID":"583b0f46-ae37-4ba6-9d33-81fd3b22f633","Type":"ContainerDied","Data":"0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.886226 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" event={"ID":"583b0f46-ae37-4ba6-9d33-81fd3b22f633","Type":"ContainerDied","Data":"b6896109eee3c66b695e19118c39ed882471cbc8942ff73cd10c613b494e40d7"} Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.886323 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s4v4x" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.898053 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-catalog-content\") pod \"3bf4f130-658a-40c0-97c5-243d77950128\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.898217 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-utilities\") pod \"3bf4f130-658a-40c0-97c5-243d77950128\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.898308 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86m7q\" (UniqueName: \"kubernetes.io/projected/3bf4f130-658a-40c0-97c5-243d77950128-kube-api-access-86m7q\") pod \"3bf4f130-658a-40c0-97c5-243d77950128\" (UID: \"3bf4f130-658a-40c0-97c5-243d77950128\") " Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.899759 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-utilities" (OuterVolumeSpecName: "utilities") pod "3bf4f130-658a-40c0-97c5-243d77950128" (UID: "3bf4f130-658a-40c0-97c5-243d77950128"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.917774 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bf4f130-658a-40c0-97c5-243d77950128-kube-api-access-86m7q" (OuterVolumeSpecName: "kube-api-access-86m7q") pod "3bf4f130-658a-40c0-97c5-243d77950128" (UID: "3bf4f130-658a-40c0-97c5-243d77950128"). InnerVolumeSpecName "kube-api-access-86m7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.919206 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s4v4x"] Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.924508 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s4v4x"] Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.939085 4758 scope.go:117] "RemoveContainer" containerID="c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.951575 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bf4f130-658a-40c0-97c5-243d77950128" (UID: "3bf4f130-658a-40c0-97c5-243d77950128"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:30 crc kubenswrapper[4758]: I1004 10:54:30.980257 4758 scope.go:117] "RemoveContainer" containerID="3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.004047 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86m7q\" (UniqueName: \"kubernetes.io/projected/3bf4f130-658a-40c0-97c5-243d77950128-kube-api-access-86m7q\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.004073 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.004083 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bf4f130-658a-40c0-97c5-243d77950128-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.014635 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.030639 4758 scope.go:117] "RemoveContainer" containerID="697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18" Oct 04 10:54:31 crc kubenswrapper[4758]: E1004 10:54:31.032847 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18\": container with ID starting with 697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18 not found: ID does not exist" containerID="697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.032893 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18"} err="failed to get container status \"697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18\": rpc error: code = NotFound desc = could not find container \"697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18\": container with ID starting with 697e29c8c9c4d934f643ba39b9c397a15a108e27e98f5ac9a6953341cb57fe18 not found: ID does not exist" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.032916 4758 scope.go:117] "RemoveContainer" containerID="c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446" Oct 04 10:54:31 crc kubenswrapper[4758]: E1004 10:54:31.033762 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446\": container with ID starting with c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446 not found: ID does not exist" containerID="c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.033795 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446"} err="failed to get container status \"c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446\": rpc error: code = NotFound desc = could not find container \"c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446\": container with ID starting with c388c82f29510b1531a1dab8700c0d2ba43c7fe3d1160b83e950671a932e4446 not found: ID does not exist" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.033822 4758 scope.go:117] "RemoveContainer" containerID="3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0" Oct 04 10:54:31 crc kubenswrapper[4758]: E1004 10:54:31.034815 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0\": container with ID starting with 3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0 not found: ID does not exist" containerID="3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.034840 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0"} err="failed to get container status \"3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0\": rpc error: code = NotFound desc = could not find container \"3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0\": container with ID starting with 3e8cd5226f751a535ef3661e00c6d716237b180273ce3d0137179f547268b8b0 not found: ID does not exist" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.034860 4758 scope.go:117] "RemoveContainer" containerID="0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.061226 4758 scope.go:117] "RemoveContainer" containerID="0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21" Oct 04 10:54:31 crc kubenswrapper[4758]: E1004 10:54:31.062425 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21\": container with ID starting with 0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21 not found: ID does not exist" containerID="0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.062459 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21"} err="failed to get container status \"0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21\": rpc error: code = NotFound desc = could not find container \"0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21\": container with ID starting with 0197050185fafde2ce66e2cf29dd4f483bae001cd03647ce61859a2c7af77d21 not found: ID does not exist" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.077052 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.104621 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn5t6\" (UniqueName: \"kubernetes.io/projected/447f9eb2-901e-42c0-813c-f03ac5f02916-kube-api-access-vn5t6\") pod \"447f9eb2-901e-42c0-813c-f03ac5f02916\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.104668 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-catalog-content\") pod \"447f9eb2-901e-42c0-813c-f03ac5f02916\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.104707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-utilities\") pod \"447f9eb2-901e-42c0-813c-f03ac5f02916\" (UID: \"447f9eb2-901e-42c0-813c-f03ac5f02916\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.105856 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-utilities" (OuterVolumeSpecName: "utilities") pod "447f9eb2-901e-42c0-813c-f03ac5f02916" (UID: "447f9eb2-901e-42c0-813c-f03ac5f02916"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.108906 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447f9eb2-901e-42c0-813c-f03ac5f02916-kube-api-access-vn5t6" (OuterVolumeSpecName: "kube-api-access-vn5t6") pod "447f9eb2-901e-42c0-813c-f03ac5f02916" (UID: "447f9eb2-901e-42c0-813c-f03ac5f02916"). InnerVolumeSpecName "kube-api-access-vn5t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.130536 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.180967 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "447f9eb2-901e-42c0-813c-f03ac5f02916" (UID: "447f9eb2-901e-42c0-813c-f03ac5f02916"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.206794 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-catalog-content\") pod \"688d59c7-04be-4f21-ae68-f134a6b72e66\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.207440 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-utilities\") pod \"688d59c7-04be-4f21-ae68-f134a6b72e66\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.207586 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qkmd\" (UniqueName: \"kubernetes.io/projected/096babec-37b1-4321-a5f6-b3a614a0b49e-kube-api-access-4qkmd\") pod \"096babec-37b1-4321-a5f6-b3a614a0b49e\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.207843 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-catalog-content\") pod \"096babec-37b1-4321-a5f6-b3a614a0b49e\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208089 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lprs9\" (UniqueName: \"kubernetes.io/projected/688d59c7-04be-4f21-ae68-f134a6b72e66-kube-api-access-lprs9\") pod \"688d59c7-04be-4f21-ae68-f134a6b72e66\" (UID: \"688d59c7-04be-4f21-ae68-f134a6b72e66\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208194 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-utilities" (OuterVolumeSpecName: "utilities") pod "688d59c7-04be-4f21-ae68-f134a6b72e66" (UID: "688d59c7-04be-4f21-ae68-f134a6b72e66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208274 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-utilities\") pod \"096babec-37b1-4321-a5f6-b3a614a0b49e\" (UID: \"096babec-37b1-4321-a5f6-b3a614a0b49e\") " Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208715 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn5t6\" (UniqueName: \"kubernetes.io/projected/447f9eb2-901e-42c0-813c-f03ac5f02916-kube-api-access-vn5t6\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208733 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208771 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447f9eb2-901e-42c0-813c-f03ac5f02916-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.208781 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.209124 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-utilities" (OuterVolumeSpecName: "utilities") pod "096babec-37b1-4321-a5f6-b3a614a0b49e" (UID: "096babec-37b1-4321-a5f6-b3a614a0b49e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.210151 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096babec-37b1-4321-a5f6-b3a614a0b49e-kube-api-access-4qkmd" (OuterVolumeSpecName: "kube-api-access-4qkmd") pod "096babec-37b1-4321-a5f6-b3a614a0b49e" (UID: "096babec-37b1-4321-a5f6-b3a614a0b49e"). InnerVolumeSpecName "kube-api-access-4qkmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.223050 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z5ck8"] Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.225335 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z5ck8"] Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.230314 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/688d59c7-04be-4f21-ae68-f134a6b72e66-kube-api-access-lprs9" (OuterVolumeSpecName: "kube-api-access-lprs9") pod "688d59c7-04be-4f21-ae68-f134a6b72e66" (UID: "688d59c7-04be-4f21-ae68-f134a6b72e66"). InnerVolumeSpecName "kube-api-access-lprs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.236518 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "096babec-37b1-4321-a5f6-b3a614a0b49e" (UID: "096babec-37b1-4321-a5f6-b3a614a0b49e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.310074 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qkmd\" (UniqueName: \"kubernetes.io/projected/096babec-37b1-4321-a5f6-b3a614a0b49e-kube-api-access-4qkmd\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.310122 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.310131 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lprs9\" (UniqueName: \"kubernetes.io/projected/688d59c7-04be-4f21-ae68-f134a6b72e66-kube-api-access-lprs9\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.310140 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/096babec-37b1-4321-a5f6-b3a614a0b49e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.316297 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "688d59c7-04be-4f21-ae68-f134a6b72e66" (UID: "688d59c7-04be-4f21-ae68-f134a6b72e66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.334184 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bf4f130-658a-40c0-97c5-243d77950128" path="/var/lib/kubelet/pods/3bf4f130-658a-40c0-97c5-243d77950128/volumes" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.334820 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" path="/var/lib/kubelet/pods/583b0f46-ae37-4ba6-9d33-81fd3b22f633/volumes" Oct 04 10:54:31 crc kubenswrapper[4758]: I1004 10:54:31.411535 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/688d59c7-04be-4f21-ae68-f134a6b72e66-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.001013 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xncn" event={"ID":"447f9eb2-901e-42c0-813c-f03ac5f02916","Type":"ContainerDied","Data":"967a467e90f214e5e95f7d5dc123bc5dd172b9710505d22bc135e51432923c0d"} Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.001118 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xncn" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.001408 4758 scope.go:117] "RemoveContainer" containerID="fa78650332def9d0743ce974624f3669282a1e8b6e290ff6d29d6505c9a4a317" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.005079 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xmk5z" event={"ID":"096babec-37b1-4321-a5f6-b3a614a0b49e","Type":"ContainerDied","Data":"9d0e04431c4966e94ddbc0d92fd1ac9b37efe7989c977b20f0d36b112519d9c6"} Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.005282 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xmk5z" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.014172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qgnlj" event={"ID":"688d59c7-04be-4f21-ae68-f134a6b72e66","Type":"ContainerDied","Data":"524d5a26e82daf35a537f324b141810575639bf7acb34ee54972fa7c5f39f1ae"} Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.014258 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qgnlj" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.021679 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qhg7v" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.025767 4758 scope.go:117] "RemoveContainer" containerID="85e03319a32408086040acfa1d5949cce0edef585cf35f47d9349077c7a2947d" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.039462 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xmk5z"] Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.044699 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xmk5z"] Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.046071 4758 scope.go:117] "RemoveContainer" containerID="cb76b8fae751853dfdaf60c47c8b813c1315c3bc332d24e5a86222e3f998cc6f" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.054455 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qgnlj"] Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.057857 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qgnlj"] Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.063518 4758 scope.go:117] "RemoveContainer" containerID="d6a20bc6d05c4d90f67d720e3f3f8b8e588aefad37402e67a59eebbd3dd90583" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.075220 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6xncn"] Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.078629 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6xncn"] Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.083383 4758 scope.go:117] "RemoveContainer" containerID="728a51e61c6079f6fe1a0a4a87b6024f3ff968117b43da605d62afb3d236959c" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.105730 4758 scope.go:117] "RemoveContainer" containerID="ed457ff81c2c02f1393fb3263c4695afb40580b6dc2a415cce5c1ff1c7c18743" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.131377 4758 scope.go:117] "RemoveContainer" containerID="8b1057789c51acaa9dc59c7af3bdc131bee8f3136d77ca30bb8610340fd5ac06" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.142567 4758 scope.go:117] "RemoveContainer" containerID="2f152fd4d185fe0dc19780c7cf55d13f26c412aede9a424fae1b71c65e9839ad" Oct 04 10:54:32 crc kubenswrapper[4758]: I1004 10:54:32.155342 4758 scope.go:117] "RemoveContainer" containerID="cdc76eb5197605a1ad0d4bc26a3f8a7adec748a2b649eb15e146e970c02c8a26" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.331766 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" path="/var/lib/kubelet/pods/096babec-37b1-4321-a5f6-b3a614a0b49e/volumes" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.332582 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" path="/var/lib/kubelet/pods/447f9eb2-901e-42c0-813c-f03ac5f02916/volumes" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.333157 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" path="/var/lib/kubelet/pods/688d59c7-04be-4f21-ae68-f134a6b72e66/volumes" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.558035 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kkfkc"] Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560195 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.560271 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560328 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.560429 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560500 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.560555 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560616 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.560669 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560733 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.560794 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560849 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.560907 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.560973 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561025 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.561085 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561165 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.561226 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561285 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.561344 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561396 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.561455 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561514 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="extract-content" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.561577 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561631 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: E1004 10:54:33.561691 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561752 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="extract-utilities" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.561994 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="688d59c7-04be-4f21-ae68-f134a6b72e66" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.562068 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="583b0f46-ae37-4ba6-9d33-81fd3b22f633" containerName="marketplace-operator" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.562157 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="447f9eb2-901e-42c0-813c-f03ac5f02916" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.562222 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="096babec-37b1-4321-a5f6-b3a614a0b49e" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.562516 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bf4f130-658a-40c0-97c5-243d77950128" containerName="registry-server" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.563718 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.569694 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kkfkc"] Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.570709 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.644362 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-catalog-content\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.644645 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-utilities\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.644744 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g8xg\" (UniqueName: \"kubernetes.io/projected/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-kube-api-access-7g8xg\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.745548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-utilities\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.745837 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g8xg\" (UniqueName: \"kubernetes.io/projected/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-kube-api-access-7g8xg\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.745964 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-catalog-content\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.746077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-utilities\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.746430 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-catalog-content\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.764136 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g8xg\" (UniqueName: \"kubernetes.io/projected/deb271af-edfd-4c2f-ae2b-1a4cbeb64f43-kube-api-access-7g8xg\") pod \"redhat-operators-kkfkc\" (UID: \"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43\") " pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:33 crc kubenswrapper[4758]: I1004 10:54:33.892699 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.150120 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s2xx8"] Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.151358 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.153066 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.158745 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2xx8"] Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.252476 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2t7h\" (UniqueName: \"kubernetes.io/projected/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-kube-api-access-m2t7h\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.252532 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-catalog-content\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.252552 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-utilities\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.312043 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kkfkc"] Oct 04 10:54:34 crc kubenswrapper[4758]: W1004 10:54:34.318641 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeb271af_edfd_4c2f_ae2b_1a4cbeb64f43.slice/crio-81d015a04275db4c7fe0cf0896b20bd62b61bb1e28a4b147b816a2b3f4b37c50 WatchSource:0}: Error finding container 81d015a04275db4c7fe0cf0896b20bd62b61bb1e28a4b147b816a2b3f4b37c50: Status 404 returned error can't find the container with id 81d015a04275db4c7fe0cf0896b20bd62b61bb1e28a4b147b816a2b3f4b37c50 Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.354090 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2t7h\" (UniqueName: \"kubernetes.io/projected/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-kube-api-access-m2t7h\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.355965 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-catalog-content\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.356148 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-utilities\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.356245 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-catalog-content\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.356694 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-utilities\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.372185 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2t7h\" (UniqueName: \"kubernetes.io/projected/7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75-kube-api-access-m2t7h\") pod \"community-operators-s2xx8\" (UID: \"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75\") " pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.521405 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:34 crc kubenswrapper[4758]: I1004 10:54:34.713964 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s2xx8"] Oct 04 10:54:34 crc kubenswrapper[4758]: W1004 10:54:34.719528 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bf770f2_a1f9_4b2e_80a9_8f6bd13b3d75.slice/crio-3626aef2b32917e73c1d89a0bd3e7c4003dfec655ceb8461906502d201c0eda3 WatchSource:0}: Error finding container 3626aef2b32917e73c1d89a0bd3e7c4003dfec655ceb8461906502d201c0eda3: Status 404 returned error can't find the container with id 3626aef2b32917e73c1d89a0bd3e7c4003dfec655ceb8461906502d201c0eda3 Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.037905 4758 generic.go:334] "Generic (PLEG): container finished" podID="deb271af-edfd-4c2f-ae2b-1a4cbeb64f43" containerID="d3e7a0b7d51471892cc69db8ae63454f66b458660cb2e30df2fedf6e42cb048b" exitCode=0 Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.037960 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkfkc" event={"ID":"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43","Type":"ContainerDied","Data":"d3e7a0b7d51471892cc69db8ae63454f66b458660cb2e30df2fedf6e42cb048b"} Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.038025 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkfkc" event={"ID":"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43","Type":"ContainerStarted","Data":"81d015a04275db4c7fe0cf0896b20bd62b61bb1e28a4b147b816a2b3f4b37c50"} Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.039214 4758 generic.go:334] "Generic (PLEG): container finished" podID="7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75" containerID="4255d0fd1dc4f3623df1fa321180e38027736d226f8f1c266bb71ba398495ccb" exitCode=0 Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.039251 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2xx8" event={"ID":"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75","Type":"ContainerDied","Data":"4255d0fd1dc4f3623df1fa321180e38027736d226f8f1c266bb71ba398495ccb"} Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.039280 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2xx8" event={"ID":"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75","Type":"ContainerStarted","Data":"3626aef2b32917e73c1d89a0bd3e7c4003dfec655ceb8461906502d201c0eda3"} Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.955180 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-49qtp"] Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.956502 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.958242 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 10:54:35 crc kubenswrapper[4758]: I1004 10:54:35.963563 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49qtp"] Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.044621 4758 generic.go:334] "Generic (PLEG): container finished" podID="7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75" containerID="7337fd1be736f3a3031c76b6ea7efffab32a5c6d2dfd739e4fc83282f00e063f" exitCode=0 Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.044654 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2xx8" event={"ID":"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75","Type":"ContainerDied","Data":"7337fd1be736f3a3031c76b6ea7efffab32a5c6d2dfd739e4fc83282f00e063f"} Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.080127 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxwf6\" (UniqueName: \"kubernetes.io/projected/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-kube-api-access-gxwf6\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.080233 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-utilities\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.080254 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-catalog-content\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.181631 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxwf6\" (UniqueName: \"kubernetes.io/projected/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-kube-api-access-gxwf6\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.181730 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-utilities\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.181754 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-catalog-content\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.182166 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-catalog-content\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.182319 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-utilities\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.207958 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxwf6\" (UniqueName: \"kubernetes.io/projected/e8129db1-3b11-4a0b-b6b0-2e019c7f45ad-kube-api-access-gxwf6\") pod \"certified-operators-49qtp\" (UID: \"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad\") " pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.308712 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.553690 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7dwq8"] Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.555351 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.563163 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.567531 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dwq8"] Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.685745 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rglsr\" (UniqueName: \"kubernetes.io/projected/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-kube-api-access-rglsr\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.685788 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-catalog-content\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.685815 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-utilities\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.739993 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49qtp"] Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.787346 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rglsr\" (UniqueName: \"kubernetes.io/projected/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-kube-api-access-rglsr\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.787655 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-catalog-content\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.787683 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-utilities\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.788122 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-utilities\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.789286 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-catalog-content\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.805721 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rglsr\" (UniqueName: \"kubernetes.io/projected/ef2a51a8-cd75-4788-93e1-5f05efa5b0b3-kube-api-access-rglsr\") pod \"redhat-marketplace-7dwq8\" (UID: \"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3\") " pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:36 crc kubenswrapper[4758]: I1004 10:54:36.907002 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.051870 4758 generic.go:334] "Generic (PLEG): container finished" podID="e8129db1-3b11-4a0b-b6b0-2e019c7f45ad" containerID="df866178ead14a50615d8faffd7ba3c7591d5ee1f3eb1c42b5dd1f84d1d26dfa" exitCode=0 Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.052286 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49qtp" event={"ID":"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad","Type":"ContainerDied","Data":"df866178ead14a50615d8faffd7ba3c7591d5ee1f3eb1c42b5dd1f84d1d26dfa"} Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.052322 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49qtp" event={"ID":"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad","Type":"ContainerStarted","Data":"935984929c8138ed3f0bc3289e6b8dbdd11ca9a22f0fe9847284c1243fa6e6d4"} Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.067253 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s2xx8" event={"ID":"7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75","Type":"ContainerStarted","Data":"1b2b4f3b1a7707f82eefe785dded05d8f973d202e84296d0a8e99c0319d1d785"} Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.090278 4758 generic.go:334] "Generic (PLEG): container finished" podID="deb271af-edfd-4c2f-ae2b-1a4cbeb64f43" containerID="f658dfb61c978231d9f2a6a4ff4d2af0e11ac739402b1f104443847fbf9c0c5b" exitCode=0 Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.090320 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkfkc" event={"ID":"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43","Type":"ContainerDied","Data":"f658dfb61c978231d9f2a6a4ff4d2af0e11ac739402b1f104443847fbf9c0c5b"} Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.111846 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s2xx8" podStartSLOduration=1.651832653 podStartE2EDuration="3.111829558s" podCreationTimestamp="2025-10-04 10:54:34 +0000 UTC" firstStartedPulling="2025-10-04 10:54:35.040902625 +0000 UTC m=+252.333553514" lastFinishedPulling="2025-10-04 10:54:36.50089953 +0000 UTC m=+253.793550419" observedRunningTime="2025-10-04 10:54:37.109475911 +0000 UTC m=+254.402126800" watchObservedRunningTime="2025-10-04 10:54:37.111829558 +0000 UTC m=+254.404480447" Oct 04 10:54:37 crc kubenswrapper[4758]: I1004 10:54:37.320714 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7dwq8"] Oct 04 10:54:37 crc kubenswrapper[4758]: W1004 10:54:37.325285 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef2a51a8_cd75_4788_93e1_5f05efa5b0b3.slice/crio-ec8871864720dca4441c20bdb55aaf0422e2b6524ac41eb9745c2be9c8258e63 WatchSource:0}: Error finding container ec8871864720dca4441c20bdb55aaf0422e2b6524ac41eb9745c2be9c8258e63: Status 404 returned error can't find the container with id ec8871864720dca4441c20bdb55aaf0422e2b6524ac41eb9745c2be9c8258e63 Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.098733 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kkfkc" event={"ID":"deb271af-edfd-4c2f-ae2b-1a4cbeb64f43","Type":"ContainerStarted","Data":"bc25c34093c2021297407b1a7a20e1445a1164d83ac5f0a28cdf4ac879001ce3"} Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.100908 4758 generic.go:334] "Generic (PLEG): container finished" podID="e8129db1-3b11-4a0b-b6b0-2e019c7f45ad" containerID="21ef5ce8f65ccc29cd9bfde9a2f457b7d316fe41b55f40b233a5d0840d619a97" exitCode=0 Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.100982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49qtp" event={"ID":"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad","Type":"ContainerDied","Data":"21ef5ce8f65ccc29cd9bfde9a2f457b7d316fe41b55f40b233a5d0840d619a97"} Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.102278 4758 generic.go:334] "Generic (PLEG): container finished" podID="ef2a51a8-cd75-4788-93e1-5f05efa5b0b3" containerID="a3f8d9e6af02297d228bd1a72c73b76c51b615a0074b01e3f9e8b8ccfbb3b0bd" exitCode=0 Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.102354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dwq8" event={"ID":"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3","Type":"ContainerDied","Data":"a3f8d9e6af02297d228bd1a72c73b76c51b615a0074b01e3f9e8b8ccfbb3b0bd"} Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.102374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dwq8" event={"ID":"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3","Type":"ContainerStarted","Data":"ec8871864720dca4441c20bdb55aaf0422e2b6524ac41eb9745c2be9c8258e63"} Oct 04 10:54:38 crc kubenswrapper[4758]: I1004 10:54:38.117311 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kkfkc" podStartSLOduration=2.579839345 podStartE2EDuration="5.117296279s" podCreationTimestamp="2025-10-04 10:54:33 +0000 UTC" firstStartedPulling="2025-10-04 10:54:35.040026666 +0000 UTC m=+252.332677565" lastFinishedPulling="2025-10-04 10:54:37.57748361 +0000 UTC m=+254.870134499" observedRunningTime="2025-10-04 10:54:38.116842395 +0000 UTC m=+255.409493284" watchObservedRunningTime="2025-10-04 10:54:38.117296279 +0000 UTC m=+255.409947168" Oct 04 10:54:40 crc kubenswrapper[4758]: I1004 10:54:40.395695 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4k6bj" Oct 04 10:54:40 crc kubenswrapper[4758]: I1004 10:54:40.447934 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gr62g"] Oct 04 10:54:41 crc kubenswrapper[4758]: I1004 10:54:41.140494 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49qtp" event={"ID":"e8129db1-3b11-4a0b-b6b0-2e019c7f45ad","Type":"ContainerStarted","Data":"615e18a9a8a2949f92bcfd7b2237e21855bf4177fe0f6a0f8e4c4ebfbe0ea83a"} Oct 04 10:54:41 crc kubenswrapper[4758]: I1004 10:54:41.142721 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dwq8" event={"ID":"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3","Type":"ContainerDied","Data":"adc6d028692e6c4fffbe82c7e424a01d2ed1d2b00460dd499eec43f8630bc8a6"} Oct 04 10:54:41 crc kubenswrapper[4758]: I1004 10:54:41.142616 4758 generic.go:334] "Generic (PLEG): container finished" podID="ef2a51a8-cd75-4788-93e1-5f05efa5b0b3" containerID="adc6d028692e6c4fffbe82c7e424a01d2ed1d2b00460dd499eec43f8630bc8a6" exitCode=0 Oct 04 10:54:41 crc kubenswrapper[4758]: I1004 10:54:41.158150 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-49qtp" podStartSLOduration=3.082116983 podStartE2EDuration="6.158122264s" podCreationTimestamp="2025-10-04 10:54:35 +0000 UTC" firstStartedPulling="2025-10-04 10:54:37.055356102 +0000 UTC m=+254.348006991" lastFinishedPulling="2025-10-04 10:54:40.131361383 +0000 UTC m=+257.424012272" observedRunningTime="2025-10-04 10:54:41.154087782 +0000 UTC m=+258.446738671" watchObservedRunningTime="2025-10-04 10:54:41.158122264 +0000 UTC m=+258.450773153" Oct 04 10:54:43 crc kubenswrapper[4758]: I1004 10:54:43.893244 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:43 crc kubenswrapper[4758]: I1004 10:54:43.894642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:43 crc kubenswrapper[4758]: I1004 10:54:43.937514 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:44 crc kubenswrapper[4758]: I1004 10:54:44.158151 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7dwq8" event={"ID":"ef2a51a8-cd75-4788-93e1-5f05efa5b0b3","Type":"ContainerStarted","Data":"e957be77c7661176b40f6d0833185f0e087b2728afe603f323127141055c7206"} Oct 04 10:54:44 crc kubenswrapper[4758]: I1004 10:54:44.207112 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kkfkc" Oct 04 10:54:44 crc kubenswrapper[4758]: I1004 10:54:44.522418 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:44 crc kubenswrapper[4758]: I1004 10:54:44.522620 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:44 crc kubenswrapper[4758]: I1004 10:54:44.564806 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:45 crc kubenswrapper[4758]: I1004 10:54:45.208963 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s2xx8" Oct 04 10:54:45 crc kubenswrapper[4758]: I1004 10:54:45.226679 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7dwq8" podStartSLOduration=4.460783989 podStartE2EDuration="9.226662354s" podCreationTimestamp="2025-10-04 10:54:36 +0000 UTC" firstStartedPulling="2025-10-04 10:54:38.103674072 +0000 UTC m=+255.396324961" lastFinishedPulling="2025-10-04 10:54:42.869552437 +0000 UTC m=+260.162203326" observedRunningTime="2025-10-04 10:54:45.184835759 +0000 UTC m=+262.477486648" watchObservedRunningTime="2025-10-04 10:54:45.226662354 +0000 UTC m=+262.519313243" Oct 04 10:54:46 crc kubenswrapper[4758]: I1004 10:54:46.309113 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:46 crc kubenswrapper[4758]: I1004 10:54:46.309415 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:46 crc kubenswrapper[4758]: I1004 10:54:46.349802 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:46 crc kubenswrapper[4758]: I1004 10:54:46.907333 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:46 crc kubenswrapper[4758]: I1004 10:54:46.907379 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:46 crc kubenswrapper[4758]: I1004 10:54:46.967427 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:47 crc kubenswrapper[4758]: I1004 10:54:47.242912 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-49qtp" Oct 04 10:54:55 crc kubenswrapper[4758]: I1004 10:54:55.132946 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" podUID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" containerName="oauth-openshift" containerID="cri-o://eb6080f3c55f04d17a08636d307b7a3d36f2e2a2ea4bd74f98d45150fdf45cc8" gracePeriod=15 Oct 04 10:54:55 crc kubenswrapper[4758]: I1004 10:54:55.814765 4758 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8zmq8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Oct 04 10:54:55 crc kubenswrapper[4758]: I1004 10:54:55.814893 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" podUID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Oct 04 10:54:55 crc kubenswrapper[4758]: E1004 10:54:55.926629 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode43ef621_3b5e_4091_ba7b_ab55ea8d5b4d.slice/crio-conmon-eb6080f3c55f04d17a08636d307b7a3d36f2e2a2ea4bd74f98d45150fdf45cc8.scope\": RecentStats: unable to find data in memory cache]" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.229745 4758 generic.go:334] "Generic (PLEG): container finished" podID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" containerID="eb6080f3c55f04d17a08636d307b7a3d36f2e2a2ea4bd74f98d45150fdf45cc8" exitCode=0 Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.229822 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" event={"ID":"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d","Type":"ContainerDied","Data":"eb6080f3c55f04d17a08636d307b7a3d36f2e2a2ea4bd74f98d45150fdf45cc8"} Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.845900 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.884747 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9"] Oct 04 10:54:56 crc kubenswrapper[4758]: E1004 10:54:56.885010 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" containerName="oauth-openshift" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.885027 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" containerName="oauth-openshift" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.885199 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" containerName="oauth-openshift" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.885682 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.909431 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9"] Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.956381 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7dwq8" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991431 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-cliconfig\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991478 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-router-certs\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991519 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-idp-0-file-data\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991560 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-trusted-ca-bundle\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-serving-cert\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991652 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-ocp-branding-template\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991686 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-service-ca\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-login\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991736 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-error\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-policies\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991779 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx6cf\" (UniqueName: \"kubernetes.io/projected/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-kube-api-access-vx6cf\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991802 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-session\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991828 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-dir\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.991852 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-provider-selection\") pod \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\" (UID: \"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d\") " Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992021 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992062 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af3bfd55-3a82-4838-abcd-0a8f76f61a28-audit-dir\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992086 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992156 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992163 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992210 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qcvx\" (UniqueName: \"kubernetes.io/projected/af3bfd55-3a82-4838-abcd-0a8f76f61a28-kube-api-access-7qcvx\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992243 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992270 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992292 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-audit-policies\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992465 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-session\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992519 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992560 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992599 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992642 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992694 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992705 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992841 4758 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992856 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.992884 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.994357 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.994780 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.997173 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.997601 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.999274 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.999525 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-kube-api-access-vx6cf" (OuterVolumeSpecName: "kube-api-access-vx6cf") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "kube-api-access-vx6cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[4758]: I1004 10:54:56.999812 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:56.999915 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.000395 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.000661 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.000802 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" (UID: "e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093514 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093581 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093610 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af3bfd55-3a82-4838-abcd-0a8f76f61a28-audit-dir\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093628 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093645 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093669 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qcvx\" (UniqueName: \"kubernetes.io/projected/af3bfd55-3a82-4838-abcd-0a8f76f61a28-kube-api-access-7qcvx\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093690 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093707 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093725 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-audit-policies\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093747 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-session\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093765 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093784 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093801 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093823 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093860 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093870 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093880 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093890 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093900 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093910 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx6cf\" (UniqueName: \"kubernetes.io/projected/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-kube-api-access-vx6cf\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093919 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093928 4758 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093936 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093945 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093955 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.093965 4758 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.094692 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.094781 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-audit-policies\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.095537 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.095575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af3bfd55-3a82-4838-abcd-0a8f76f61a28-audit-dir\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.096301 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.097302 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.097768 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.098629 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.098886 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-session\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.099020 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.101154 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.102289 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.106230 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af3bfd55-3a82-4838-abcd-0a8f76f61a28-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.113619 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qcvx\" (UniqueName: \"kubernetes.io/projected/af3bfd55-3a82-4838-abcd-0a8f76f61a28-kube-api-access-7qcvx\") pod \"oauth-openshift-5d4f55d7c5-9bkd9\" (UID: \"af3bfd55-3a82-4838-abcd-0a8f76f61a28\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.214793 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.237657 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" event={"ID":"e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d","Type":"ContainerDied","Data":"487ee9c22b47491629b4e755e88a0fa1fb78199296ff64a7ffbaac9c2891b528"} Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.237689 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8zmq8" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.237712 4758 scope.go:117] "RemoveContainer" containerID="eb6080f3c55f04d17a08636d307b7a3d36f2e2a2ea4bd74f98d45150fdf45cc8" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.263653 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8zmq8"] Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.268469 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8zmq8"] Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.344597 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d" path="/var/lib/kubelet/pods/e43ef621-3b5e-4091-ba7b-ab55ea8d5b4d/volumes" Oct 04 10:54:57 crc kubenswrapper[4758]: I1004 10:54:57.635992 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9"] Oct 04 10:54:57 crc kubenswrapper[4758]: W1004 10:54:57.643017 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf3bfd55_3a82_4838_abcd_0a8f76f61a28.slice/crio-4b6055568adee403ff353fb03de0d4085232a43d35f60cdb0204d2f573ec856b WatchSource:0}: Error finding container 4b6055568adee403ff353fb03de0d4085232a43d35f60cdb0204d2f573ec856b: Status 404 returned error can't find the container with id 4b6055568adee403ff353fb03de0d4085232a43d35f60cdb0204d2f573ec856b Oct 04 10:54:58 crc kubenswrapper[4758]: I1004 10:54:58.250178 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" event={"ID":"af3bfd55-3a82-4838-abcd-0a8f76f61a28","Type":"ContainerStarted","Data":"8ca35c61c5112549d128a8acf14e07052586394a9718697d3844d9e1c0e72c80"} Oct 04 10:54:58 crc kubenswrapper[4758]: I1004 10:54:58.250728 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" event={"ID":"af3bfd55-3a82-4838-abcd-0a8f76f61a28","Type":"ContainerStarted","Data":"4b6055568adee403ff353fb03de0d4085232a43d35f60cdb0204d2f573ec856b"} Oct 04 10:54:58 crc kubenswrapper[4758]: I1004 10:54:58.252129 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:58 crc kubenswrapper[4758]: I1004 10:54:58.554943 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" Oct 04 10:54:58 crc kubenswrapper[4758]: I1004 10:54:58.574831 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-9bkd9" podStartSLOduration=28.574814775 podStartE2EDuration="28.574814775s" podCreationTimestamp="2025-10-04 10:54:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:58.286351323 +0000 UTC m=+275.579002212" watchObservedRunningTime="2025-10-04 10:54:58.574814775 +0000 UTC m=+275.867465664" Oct 04 10:55:05 crc kubenswrapper[4758]: I1004 10:55:05.479922 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" podUID="dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" containerName="registry" containerID="cri-o://7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99" gracePeriod=30 Oct 04 10:55:05 crc kubenswrapper[4758]: I1004 10:55:05.954415 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.127031 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-bound-sa-token\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.127449 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-ca-trust-extracted\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.127578 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-trusted-ca\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.127705 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-certificates\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.127936 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.128084 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8km7\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-kube-api-access-c8km7\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.128249 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-installation-pull-secrets\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.128351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-tls\") pod \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\" (UID: \"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a\") " Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.128568 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.128769 4758 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.129754 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.132947 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.133371 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.136333 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-kube-api-access-c8km7" (OuterVolumeSpecName: "kube-api-access-c8km7") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "kube-api-access-c8km7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.140199 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.142048 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.149552 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" (UID: "dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.230444 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8km7\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-kube-api-access-c8km7\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.230497 4758 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.230517 4758 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.230539 4758 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.230557 4758 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.230573 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.306509 4758 generic.go:334] "Generic (PLEG): container finished" podID="dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" containerID="7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99" exitCode=0 Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.306599 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.307086 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" event={"ID":"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a","Type":"ContainerDied","Data":"7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99"} Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.307423 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gr62g" event={"ID":"dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a","Type":"ContainerDied","Data":"a11e8b59ab8f3465c28f2e6620e108ff0a95f1e40f413d8d4a93915c3fb43b44"} Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.307510 4758 scope.go:117] "RemoveContainer" containerID="7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.334711 4758 scope.go:117] "RemoveContainer" containerID="7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99" Oct 04 10:55:06 crc kubenswrapper[4758]: E1004 10:55:06.335641 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99\": container with ID starting with 7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99 not found: ID does not exist" containerID="7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.335702 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99"} err="failed to get container status \"7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99\": rpc error: code = NotFound desc = could not find container \"7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99\": container with ID starting with 7072a60788ffc87fc8eaac119e63739c9e034d6e21183077e96c5cc8bba56f99 not found: ID does not exist" Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.359137 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gr62g"] Oct 04 10:55:06 crc kubenswrapper[4758]: I1004 10:55:06.371188 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gr62g"] Oct 04 10:55:07 crc kubenswrapper[4758]: I1004 10:55:07.337447 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" path="/var/lib/kubelet/pods/dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a/volumes" Oct 04 10:56:31 crc kubenswrapper[4758]: I1004 10:56:31.249767 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:56:31 crc kubenswrapper[4758]: I1004 10:56:31.250573 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:57:01 crc kubenswrapper[4758]: I1004 10:57:01.249740 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:57:01 crc kubenswrapper[4758]: I1004 10:57:01.250310 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:57:31 crc kubenswrapper[4758]: I1004 10:57:31.249542 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:57:31 crc kubenswrapper[4758]: I1004 10:57:31.250179 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:57:31 crc kubenswrapper[4758]: I1004 10:57:31.250230 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 10:57:31 crc kubenswrapper[4758]: I1004 10:57:31.250839 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"770a65d2316af3d496832d48f9bf7d288866132c38f07143143f0ef4b5eb238f"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:57:31 crc kubenswrapper[4758]: I1004 10:57:31.250917 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://770a65d2316af3d496832d48f9bf7d288866132c38f07143143f0ef4b5eb238f" gracePeriod=600 Oct 04 10:57:32 crc kubenswrapper[4758]: I1004 10:57:32.169569 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="770a65d2316af3d496832d48f9bf7d288866132c38f07143143f0ef4b5eb238f" exitCode=0 Oct 04 10:57:32 crc kubenswrapper[4758]: I1004 10:57:32.169648 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"770a65d2316af3d496832d48f9bf7d288866132c38f07143143f0ef4b5eb238f"} Oct 04 10:57:32 crc kubenswrapper[4758]: I1004 10:57:32.170595 4758 scope.go:117] "RemoveContainer" containerID="490cdcf5e16874fa92dbfb0ddf7ea20c91ac9d416a231c50ea41cbb02d87cc96" Oct 04 10:57:33 crc kubenswrapper[4758]: I1004 10:57:33.178059 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"4a4ba30925ab5365601b714b606e8b4bf9695f793e826fa68e869bdae0e9d640"} Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.536501 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qh9h8"] Oct 04 10:59:51 crc kubenswrapper[4758]: E1004 10:59:51.537201 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" containerName="registry" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.537213 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" containerName="registry" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.537338 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd4cef36-6bc0-4a10-8d0e-9c4fca44fd2a" containerName="registry" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.537677 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.540752 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-zr5ff" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.540950 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.542280 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.553403 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qh9h8"] Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.564068 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xmgtx"] Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.564756 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xmgtx" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.570635 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xmgtx"] Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.573387 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-9f4lh" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.579461 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fwh97"] Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.579999 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.583297 4758 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-rnmq7" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.598383 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fwh97"] Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.627852 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdmx4\" (UniqueName: \"kubernetes.io/projected/d2e6b6be-3b22-431d-8401-ce9e45bbb7f1-kube-api-access-mdmx4\") pod \"cert-manager-5b446d88c5-xmgtx\" (UID: \"d2e6b6be-3b22-431d-8401-ce9e45bbb7f1\") " pod="cert-manager/cert-manager-5b446d88c5-xmgtx" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.627913 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btpnk\" (UniqueName: \"kubernetes.io/projected/8496f61c-f11f-4358-8f0e-eda4eb9e42df-kube-api-access-btpnk\") pod \"cert-manager-cainjector-7f985d654d-qh9h8\" (UID: \"8496f61c-f11f-4358-8f0e-eda4eb9e42df\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.729072 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsqhc\" (UniqueName: \"kubernetes.io/projected/0396bbf8-fa84-4c24-916a-0c67c99f1029-kube-api-access-wsqhc\") pod \"cert-manager-webhook-5655c58dd6-fwh97\" (UID: \"0396bbf8-fa84-4c24-916a-0c67c99f1029\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.729145 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btpnk\" (UniqueName: \"kubernetes.io/projected/8496f61c-f11f-4358-8f0e-eda4eb9e42df-kube-api-access-btpnk\") pod \"cert-manager-cainjector-7f985d654d-qh9h8\" (UID: \"8496f61c-f11f-4358-8f0e-eda4eb9e42df\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.729516 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdmx4\" (UniqueName: \"kubernetes.io/projected/d2e6b6be-3b22-431d-8401-ce9e45bbb7f1-kube-api-access-mdmx4\") pod \"cert-manager-5b446d88c5-xmgtx\" (UID: \"d2e6b6be-3b22-431d-8401-ce9e45bbb7f1\") " pod="cert-manager/cert-manager-5b446d88c5-xmgtx" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.760952 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btpnk\" (UniqueName: \"kubernetes.io/projected/8496f61c-f11f-4358-8f0e-eda4eb9e42df-kube-api-access-btpnk\") pod \"cert-manager-cainjector-7f985d654d-qh9h8\" (UID: \"8496f61c-f11f-4358-8f0e-eda4eb9e42df\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.764299 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdmx4\" (UniqueName: \"kubernetes.io/projected/d2e6b6be-3b22-431d-8401-ce9e45bbb7f1-kube-api-access-mdmx4\") pod \"cert-manager-5b446d88c5-xmgtx\" (UID: \"d2e6b6be-3b22-431d-8401-ce9e45bbb7f1\") " pod="cert-manager/cert-manager-5b446d88c5-xmgtx" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.831358 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsqhc\" (UniqueName: \"kubernetes.io/projected/0396bbf8-fa84-4c24-916a-0c67c99f1029-kube-api-access-wsqhc\") pod \"cert-manager-webhook-5655c58dd6-fwh97\" (UID: \"0396bbf8-fa84-4c24-916a-0c67c99f1029\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.846813 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsqhc\" (UniqueName: \"kubernetes.io/projected/0396bbf8-fa84-4c24-916a-0c67c99f1029-kube-api-access-wsqhc\") pod \"cert-manager-webhook-5655c58dd6-fwh97\" (UID: \"0396bbf8-fa84-4c24-916a-0c67c99f1029\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.854495 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.883289 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xmgtx" Oct 04 10:59:51 crc kubenswrapper[4758]: I1004 10:59:51.893059 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 10:59:52 crc kubenswrapper[4758]: I1004 10:59:52.084377 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qh9h8"] Oct 04 10:59:52 crc kubenswrapper[4758]: I1004 10:59:52.084850 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:59:52 crc kubenswrapper[4758]: I1004 10:59:52.149576 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fwh97"] Oct 04 10:59:52 crc kubenswrapper[4758]: W1004 10:59:52.151795 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0396bbf8_fa84_4c24_916a_0c67c99f1029.slice/crio-9e4be4d1b444f5ff83fed82b1d6adf7bd34dde8f9bf533ff035bdb418a87c30b WatchSource:0}: Error finding container 9e4be4d1b444f5ff83fed82b1d6adf7bd34dde8f9bf533ff035bdb418a87c30b: Status 404 returned error can't find the container with id 9e4be4d1b444f5ff83fed82b1d6adf7bd34dde8f9bf533ff035bdb418a87c30b Oct 04 10:59:52 crc kubenswrapper[4758]: I1004 10:59:52.199070 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xmgtx"] Oct 04 10:59:53 crc kubenswrapper[4758]: I1004 10:59:53.066409 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" event={"ID":"0396bbf8-fa84-4c24-916a-0c67c99f1029","Type":"ContainerStarted","Data":"9e4be4d1b444f5ff83fed82b1d6adf7bd34dde8f9bf533ff035bdb418a87c30b"} Oct 04 10:59:53 crc kubenswrapper[4758]: I1004 10:59:53.067796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xmgtx" event={"ID":"d2e6b6be-3b22-431d-8401-ce9e45bbb7f1","Type":"ContainerStarted","Data":"a949fc4be380c0c1339612431400ad8bb68a6821b15f0e86ced54f6fca7e4798"} Oct 04 10:59:53 crc kubenswrapper[4758]: I1004 10:59:53.068696 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" event={"ID":"8496f61c-f11f-4358-8f0e-eda4eb9e42df","Type":"ContainerStarted","Data":"ab159eb6dac444fb9ffc44ccf5c8dbb0a304f94154be9e245f30167e5a56f91d"} Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.084819 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xmgtx" event={"ID":"d2e6b6be-3b22-431d-8401-ce9e45bbb7f1","Type":"ContainerStarted","Data":"85c1884f71727c268334a5f2b09a033853b59e08f804c24eb5fbbedcde424267"} Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.087689 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" event={"ID":"8496f61c-f11f-4358-8f0e-eda4eb9e42df","Type":"ContainerStarted","Data":"77698fb1fdcd59b6bbd60c99e9634b2b2c519461ad40623d04447fdf600fe34f"} Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.088849 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" event={"ID":"0396bbf8-fa84-4c24-916a-0c67c99f1029","Type":"ContainerStarted","Data":"71eaddab96af54fdab69c4685eb5c998de149eaae5ccdef79053ccb18ce4c8d2"} Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.089374 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.102724 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-xmgtx" podStartSLOduration=1.6947343670000001 podStartE2EDuration="5.102707573s" podCreationTimestamp="2025-10-04 10:59:51 +0000 UTC" firstStartedPulling="2025-10-04 10:59:52.211805945 +0000 UTC m=+569.504456834" lastFinishedPulling="2025-10-04 10:59:55.619779151 +0000 UTC m=+572.912430040" observedRunningTime="2025-10-04 10:59:56.100200143 +0000 UTC m=+573.392851032" watchObservedRunningTime="2025-10-04 10:59:56.102707573 +0000 UTC m=+573.395358462" Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.124530 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" podStartSLOduration=1.654098706 podStartE2EDuration="5.12451274s" podCreationTimestamp="2025-10-04 10:59:51 +0000 UTC" firstStartedPulling="2025-10-04 10:59:52.160183649 +0000 UTC m=+569.452834538" lastFinishedPulling="2025-10-04 10:59:55.630597683 +0000 UTC m=+572.923248572" observedRunningTime="2025-10-04 10:59:56.123766529 +0000 UTC m=+573.416417408" watchObservedRunningTime="2025-10-04 10:59:56.12451274 +0000 UTC m=+573.417163629" Oct 04 10:59:56 crc kubenswrapper[4758]: I1004 10:59:56.146986 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh9h8" podStartSLOduration=1.590804624 podStartE2EDuration="5.146969455s" podCreationTimestamp="2025-10-04 10:59:51 +0000 UTC" firstStartedPulling="2025-10-04 10:59:52.084568304 +0000 UTC m=+569.377219203" lastFinishedPulling="2025-10-04 10:59:55.640733145 +0000 UTC m=+572.933384034" observedRunningTime="2025-10-04 10:59:56.140601978 +0000 UTC m=+573.433252867" watchObservedRunningTime="2025-10-04 10:59:56.146969455 +0000 UTC m=+573.439620344" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.127522 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2"] Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.128789 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.131580 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.133423 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.135088 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2"] Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.241827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54c2c19f-863a-4faf-9818-1ce9ab1ada00-secret-volume\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.242328 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7k7x\" (UniqueName: \"kubernetes.io/projected/54c2c19f-863a-4faf-9818-1ce9ab1ada00-kube-api-access-c7k7x\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.242396 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54c2c19f-863a-4faf-9818-1ce9ab1ada00-config-volume\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.343255 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54c2c19f-863a-4faf-9818-1ce9ab1ada00-secret-volume\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.343316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7k7x\" (UniqueName: \"kubernetes.io/projected/54c2c19f-863a-4faf-9818-1ce9ab1ada00-kube-api-access-c7k7x\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.343349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54c2c19f-863a-4faf-9818-1ce9ab1ada00-config-volume\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.344717 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54c2c19f-863a-4faf-9818-1ce9ab1ada00-config-volume\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.355795 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54c2c19f-863a-4faf-9818-1ce9ab1ada00-secret-volume\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.361437 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7k7x\" (UniqueName: \"kubernetes.io/projected/54c2c19f-863a-4faf-9818-1ce9ab1ada00-kube-api-access-c7k7x\") pod \"collect-profiles-29326260-n59c2\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.454385 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:00 crc kubenswrapper[4758]: I1004 11:00:00.656476 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2"] Oct 04 11:00:01 crc kubenswrapper[4758]: I1004 11:00:01.118882 4758 generic.go:334] "Generic (PLEG): container finished" podID="54c2c19f-863a-4faf-9818-1ce9ab1ada00" containerID="a7a604075251b79e81ab1b814cb2867b52991cd45ca8cb8d33bf64f89a3440cb" exitCode=0 Oct 04 11:00:01 crc kubenswrapper[4758]: I1004 11:00:01.118931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" event={"ID":"54c2c19f-863a-4faf-9818-1ce9ab1ada00","Type":"ContainerDied","Data":"a7a604075251b79e81ab1b814cb2867b52991cd45ca8cb8d33bf64f89a3440cb"} Oct 04 11:00:01 crc kubenswrapper[4758]: I1004 11:00:01.119264 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" event={"ID":"54c2c19f-863a-4faf-9818-1ce9ab1ada00","Type":"ContainerStarted","Data":"fbf6d8bda79ada7852d55a4d0433cef8f8546e3b5c9d65ec54ebfd1565a6c1b3"} Oct 04 11:00:01 crc kubenswrapper[4758]: I1004 11:00:01.249532 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:00:01 crc kubenswrapper[4758]: I1004 11:00:01.249598 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:00:01 crc kubenswrapper[4758]: I1004 11:00:01.896496 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-fwh97" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.068991 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hbnh9"] Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.069794 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-controller" containerID="cri-o://a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.069847 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="nbdb" containerID="cri-o://80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.069890 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="northd" containerID="cri-o://18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.069960 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-node" containerID="cri-o://9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.070018 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-acl-logging" containerID="cri-o://1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.070070 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="sbdb" containerID="cri-o://1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.069862 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.110881 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" containerID="cri-o://9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" gracePeriod=30 Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.260456 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.367146 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54c2c19f-863a-4faf-9818-1ce9ab1ada00-config-volume\") pod \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.367254 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54c2c19f-863a-4faf-9818-1ce9ab1ada00-secret-volume\") pod \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.367338 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7k7x\" (UniqueName: \"kubernetes.io/projected/54c2c19f-863a-4faf-9818-1ce9ab1ada00-kube-api-access-c7k7x\") pod \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\" (UID: \"54c2c19f-863a-4faf-9818-1ce9ab1ada00\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.367827 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54c2c19f-863a-4faf-9818-1ce9ab1ada00-config-volume" (OuterVolumeSpecName: "config-volume") pod "54c2c19f-863a-4faf-9818-1ce9ab1ada00" (UID: "54c2c19f-863a-4faf-9818-1ce9ab1ada00"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.371767 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c2c19f-863a-4faf-9818-1ce9ab1ada00-kube-api-access-c7k7x" (OuterVolumeSpecName: "kube-api-access-c7k7x") pod "54c2c19f-863a-4faf-9818-1ce9ab1ada00" (UID: "54c2c19f-863a-4faf-9818-1ce9ab1ada00"). InnerVolumeSpecName "kube-api-access-c7k7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.372905 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54c2c19f-863a-4faf-9818-1ce9ab1ada00-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "54c2c19f-863a-4faf-9818-1ce9ab1ada00" (UID: "54c2c19f-863a-4faf-9818-1ce9ab1ada00"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.440446 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/4.log" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.441225 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/3.log" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.442764 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovn-acl-logging/0.log" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.443131 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovn-controller/0.log" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.443423 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.468575 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7k7x\" (UniqueName: \"kubernetes.io/projected/54c2c19f-863a-4faf-9818-1ce9ab1ada00-kube-api-access-c7k7x\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.468598 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/54c2c19f-863a-4faf-9818-1ce9ab1ada00-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.468608 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/54c2c19f-863a-4faf-9818-1ce9ab1ada00-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487681 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mtnlv"] Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487875 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487887 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487897 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="sbdb" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487903 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="sbdb" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487913 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487920 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487926 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487931 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487942 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kubecfg-setup" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487948 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kubecfg-setup" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487957 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487963 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487970 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="nbdb" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487975 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="nbdb" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.487984 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c2c19f-863a-4faf-9818-1ce9ab1ada00" containerName="collect-profiles" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.487990 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c2c19f-863a-4faf-9818-1ce9ab1ada00" containerName="collect-profiles" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.488002 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488007 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.488013 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="northd" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488018 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="northd" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.488026 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-acl-logging" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488032 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-acl-logging" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.488040 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-node" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488045 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-node" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488149 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-node" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488160 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488167 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488175 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488181 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488190 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488199 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c2c19f-863a-4faf-9818-1ce9ab1ada00" containerName="collect-profiles" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488207 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="nbdb" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488213 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovn-acl-logging" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488219 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488227 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="northd" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488233 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="sbdb" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.488315 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488322 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: E1004 11:00:02.488332 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488339 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.488422 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cec2b-782d-4033-a24c-4129a935cf27" containerName="ovnkube-controller" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.489913 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.569572 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-script-lib\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.569800 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-kubelet\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.569891 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.569998 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-systemd-units\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570035 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570200 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570336 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-node-log" (OuterVolumeSpecName: "node-log") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570432 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-node-log\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570555 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-ovn\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570668 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063cec2b-782d-4033-a24c-4129a935cf27-ovn-node-metrics-cert\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.570593 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571295 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-etc-openvswitch\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571368 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571375 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-var-lib-openvswitch\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571405 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571430 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-netd\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571498 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-ovn-kubernetes\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571494 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571521 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-openvswitch\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571544 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-env-overrides\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571586 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-var-lib-cni-networks-ovn-kubernetes\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571605 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-config\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571630 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-systemd\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571645 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-log-socket\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571665 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-netns\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571689 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-slash\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571709 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-bin\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571732 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tk2m\" (UniqueName: \"kubernetes.io/projected/063cec2b-782d-4033-a24c-4129a935cf27-kube-api-access-5tk2m\") pod \"063cec2b-782d-4033-a24c-4129a935cf27\" (UID: \"063cec2b-782d-4033-a24c-4129a935cf27\") " Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571753 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571786 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571783 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571805 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-log-socket" (OuterVolumeSpecName: "log-socket") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571825 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571849 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-slash" (OuterVolumeSpecName: "host-slash") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.571889 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572001 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-kubelet\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572044 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-run-netns\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572115 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572132 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-ovnkube-config\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572173 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-log-socket\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572210 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stcb8\" (UniqueName: \"kubernetes.io/projected/29580f20-e100-43a2-a2ad-deb27a5e0340-kube-api-access-stcb8\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572253 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-slash\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572327 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572357 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-ovn\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572376 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-ovnkube-script-lib\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572396 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-etc-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572422 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-env-overrides\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572520 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-cni-bin\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572554 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-systemd-units\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572578 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/29580f20-e100-43a2-a2ad-deb27a5e0340-ovn-node-metrics-cert\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572614 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-node-log\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572651 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-run-ovn-kubernetes\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572686 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-var-lib-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572708 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-systemd\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572750 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-cni-netd\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572800 4758 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572826 4758 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.572906 4758 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573017 4758 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573030 4758 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573045 4758 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573058 4758 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573072 4758 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573088 4758 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573115 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573131 4758 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573143 4758 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573156 4758 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573168 4758 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573180 4758 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.573208 4758 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.574543 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/063cec2b-782d-4033-a24c-4129a935cf27-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.575179 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/063cec2b-782d-4033-a24c-4129a935cf27-kube-api-access-5tk2m" (OuterVolumeSpecName: "kube-api-access-5tk2m") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "kube-api-access-5tk2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.582598 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "063cec2b-782d-4033-a24c-4129a935cf27" (UID: "063cec2b-782d-4033-a24c-4129a935cf27"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673703 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673748 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-ovn\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673767 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-ovnkube-script-lib\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673786 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-etc-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673809 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-env-overrides\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673839 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-cni-bin\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-systemd-units\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673870 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/29580f20-e100-43a2-a2ad-deb27a5e0340-ovn-node-metrics-cert\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673885 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-node-log\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673903 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-run-ovn-kubernetes\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673921 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-var-lib-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673934 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-cni-netd\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673919 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673987 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-systemd\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.673948 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-systemd\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674083 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674189 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-kubelet\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674234 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-run-netns\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674289 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-ovnkube-config\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674325 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-log-socket\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stcb8\" (UniqueName: \"kubernetes.io/projected/29580f20-e100-43a2-a2ad-deb27a5e0340-kube-api-access-stcb8\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674408 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-slash\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674515 4758 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/063cec2b-782d-4033-a24c-4129a935cf27-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674539 4758 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/063cec2b-782d-4033-a24c-4129a935cf27-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674560 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tk2m\" (UniqueName: \"kubernetes.io/projected/063cec2b-782d-4033-a24c-4129a935cf27-kube-api-access-5tk2m\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674581 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/063cec2b-782d-4033-a24c-4129a935cf27-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674623 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-env-overrides\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674629 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-slash\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674658 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-cni-bin\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674022 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-etc-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674693 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-systemd-units\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674698 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674744 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-kubelet\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674790 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-run-netns\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674790 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-ovnkube-script-lib\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674893 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-run-ovn\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674935 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-run-ovn-kubernetes\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.674971 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-node-log\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.675035 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-log-socket\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.675134 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-var-lib-openvswitch\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.675174 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/29580f20-e100-43a2-a2ad-deb27a5e0340-host-cni-netd\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.675904 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/29580f20-e100-43a2-a2ad-deb27a5e0340-ovnkube-config\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.677761 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/29580f20-e100-43a2-a2ad-deb27a5e0340-ovn-node-metrics-cert\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.690679 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stcb8\" (UniqueName: \"kubernetes.io/projected/29580f20-e100-43a2-a2ad-deb27a5e0340-kube-api-access-stcb8\") pod \"ovnkube-node-mtnlv\" (UID: \"29580f20-e100-43a2-a2ad-deb27a5e0340\") " pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: I1004 11:00:02.803618 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:02 crc kubenswrapper[4758]: W1004 11:00:02.823638 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29580f20_e100_43a2_a2ad_deb27a5e0340.slice/crio-39a9684132ee61c0ef735c4ebb4cf03992ab956a58393da775e70a3b79fdfac4 WatchSource:0}: Error finding container 39a9684132ee61c0ef735c4ebb4cf03992ab956a58393da775e70a3b79fdfac4: Status 404 returned error can't find the container with id 39a9684132ee61c0ef735c4ebb4cf03992ab956a58393da775e70a3b79fdfac4 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.132004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" event={"ID":"54c2c19f-863a-4faf-9818-1ce9ab1ada00","Type":"ContainerDied","Data":"fbf6d8bda79ada7852d55a4d0433cef8f8546e3b5c9d65ec54ebfd1565a6c1b3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.132085 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbf6d8bda79ada7852d55a4d0433cef8f8546e3b5c9d65ec54ebfd1565a6c1b3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.132224 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.134260 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/4.log" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.134944 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovnkube-controller/3.log" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.138430 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovn-acl-logging/0.log" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.138967 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hbnh9_063cec2b-782d-4033-a24c-4129a935cf27/ovn-controller/0.log" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139428 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" exitCode=2 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139450 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" exitCode=0 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139457 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" exitCode=0 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139465 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" exitCode=0 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139472 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" exitCode=0 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139479 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" exitCode=0 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139487 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" exitCode=143 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139494 4758 generic.go:334] "Generic (PLEG): container finished" podID="063cec2b-782d-4033-a24c-4129a935cf27" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" exitCode=143 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139511 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139581 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139605 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139627 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139645 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139658 4758 scope.go:117] "RemoveContainer" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139665 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139683 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139701 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139714 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139725 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139737 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139748 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139761 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139772 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139783 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139801 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139819 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139832 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139844 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139854 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139865 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139875 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139886 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139896 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139907 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139919 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139934 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139949 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139963 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139973 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139984 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.139995 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140005 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140016 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140031 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140041 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140052 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140071 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hbnh9" event={"ID":"063cec2b-782d-4033-a24c-4129a935cf27","Type":"ContainerDied","Data":"1d3131a1ae8707e55ac8d5d638f1d2e8de3048e55c1c148906fdf96d320dc1a7"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140095 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140190 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140202 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140213 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140255 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140267 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140278 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140288 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140299 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.140311 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.141340 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/2.log" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.142924 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/1.log" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.142960 4758 generic.go:334] "Generic (PLEG): container finished" podID="3875fe1d-de8c-4266-bf2b-e07c633b85dc" containerID="1570a60d9f80108e4ebf690d60a5ecffd03622708015353cc0f5247388c750f8" exitCode=2 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.143019 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerDied","Data":"1570a60d9f80108e4ebf690d60a5ecffd03622708015353cc0f5247388c750f8"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.143047 4758 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.143414 4758 scope.go:117] "RemoveContainer" containerID="1570a60d9f80108e4ebf690d60a5ecffd03622708015353cc0f5247388c750f8" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.143601 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bpvxr_openshift-multus(3875fe1d-de8c-4266-bf2b-e07c633b85dc)\"" pod="openshift-multus/multus-bpvxr" podUID="3875fe1d-de8c-4266-bf2b-e07c633b85dc" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.145597 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerDied","Data":"6648fda6be0b07ddbe4a6f49385f36d231811edc46dc0d474316f471b5c8216d"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.145593 4758 generic.go:334] "Generic (PLEG): container finished" podID="29580f20-e100-43a2-a2ad-deb27a5e0340" containerID="6648fda6be0b07ddbe4a6f49385f36d231811edc46dc0d474316f471b5c8216d" exitCode=0 Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.146865 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"39a9684132ee61c0ef735c4ebb4cf03992ab956a58393da775e70a3b79fdfac4"} Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.163646 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.207270 4758 scope.go:117] "RemoveContainer" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.223661 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hbnh9"] Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.233224 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hbnh9"] Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.241394 4758 scope.go:117] "RemoveContainer" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.259360 4758 scope.go:117] "RemoveContainer" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.272760 4758 scope.go:117] "RemoveContainer" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.299329 4758 scope.go:117] "RemoveContainer" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.320453 4758 scope.go:117] "RemoveContainer" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.331949 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="063cec2b-782d-4033-a24c-4129a935cf27" path="/var/lib/kubelet/pods/063cec2b-782d-4033-a24c-4129a935cf27/volumes" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.336301 4758 scope.go:117] "RemoveContainer" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.350076 4758 scope.go:117] "RemoveContainer" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.386377 4758 scope.go:117] "RemoveContainer" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.386784 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": container with ID starting with 9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62 not found: ID does not exist" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.386811 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} err="failed to get container status \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": rpc error: code = NotFound desc = could not find container \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": container with ID starting with 9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.386832 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.387093 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": container with ID starting with c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b not found: ID does not exist" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.387133 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} err="failed to get container status \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": rpc error: code = NotFound desc = could not find container \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": container with ID starting with c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.387146 4758 scope.go:117] "RemoveContainer" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.387509 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": container with ID starting with 1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3 not found: ID does not exist" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.387549 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} err="failed to get container status \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": rpc error: code = NotFound desc = could not find container \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": container with ID starting with 1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.387563 4758 scope.go:117] "RemoveContainer" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.387919 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": container with ID starting with 80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f not found: ID does not exist" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.387963 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} err="failed to get container status \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": rpc error: code = NotFound desc = could not find container \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": container with ID starting with 80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.387994 4758 scope.go:117] "RemoveContainer" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.388357 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": container with ID starting with 18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d not found: ID does not exist" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.388378 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} err="failed to get container status \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": rpc error: code = NotFound desc = could not find container \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": container with ID starting with 18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.388391 4758 scope.go:117] "RemoveContainer" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.388634 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": container with ID starting with ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252 not found: ID does not exist" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.388671 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} err="failed to get container status \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": rpc error: code = NotFound desc = could not find container \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": container with ID starting with ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.388684 4758 scope.go:117] "RemoveContainer" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.388927 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": container with ID starting with 9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69 not found: ID does not exist" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.388953 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} err="failed to get container status \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": rpc error: code = NotFound desc = could not find container \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": container with ID starting with 9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.388971 4758 scope.go:117] "RemoveContainer" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.389215 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": container with ID starting with 1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9 not found: ID does not exist" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.389239 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} err="failed to get container status \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": rpc error: code = NotFound desc = could not find container \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": container with ID starting with 1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.389254 4758 scope.go:117] "RemoveContainer" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.389467 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": container with ID starting with a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6 not found: ID does not exist" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.389508 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} err="failed to get container status \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": rpc error: code = NotFound desc = could not find container \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": container with ID starting with a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.389522 4758 scope.go:117] "RemoveContainer" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" Oct 04 11:00:03 crc kubenswrapper[4758]: E1004 11:00:03.389863 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": container with ID starting with 2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3 not found: ID does not exist" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.389889 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} err="failed to get container status \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": rpc error: code = NotFound desc = could not find container \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": container with ID starting with 2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.389903 4758 scope.go:117] "RemoveContainer" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.390173 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} err="failed to get container status \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": rpc error: code = NotFound desc = could not find container \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": container with ID starting with 9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.390261 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.390597 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} err="failed to get container status \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": rpc error: code = NotFound desc = could not find container \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": container with ID starting with c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.390635 4758 scope.go:117] "RemoveContainer" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.390863 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} err="failed to get container status \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": rpc error: code = NotFound desc = could not find container \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": container with ID starting with 1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.390884 4758 scope.go:117] "RemoveContainer" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391155 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} err="failed to get container status \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": rpc error: code = NotFound desc = could not find container \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": container with ID starting with 80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391174 4758 scope.go:117] "RemoveContainer" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391435 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} err="failed to get container status \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": rpc error: code = NotFound desc = could not find container \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": container with ID starting with 18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391453 4758 scope.go:117] "RemoveContainer" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391759 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} err="failed to get container status \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": rpc error: code = NotFound desc = could not find container \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": container with ID starting with ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391779 4758 scope.go:117] "RemoveContainer" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.391986 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} err="failed to get container status \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": rpc error: code = NotFound desc = could not find container \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": container with ID starting with 9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392044 4758 scope.go:117] "RemoveContainer" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392286 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} err="failed to get container status \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": rpc error: code = NotFound desc = could not find container \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": container with ID starting with 1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392318 4758 scope.go:117] "RemoveContainer" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392521 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} err="failed to get container status \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": rpc error: code = NotFound desc = could not find container \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": container with ID starting with a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392539 4758 scope.go:117] "RemoveContainer" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392761 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} err="failed to get container status \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": rpc error: code = NotFound desc = could not find container \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": container with ID starting with 2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392777 4758 scope.go:117] "RemoveContainer" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392964 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} err="failed to get container status \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": rpc error: code = NotFound desc = could not find container \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": container with ID starting with 9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.392990 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.393193 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} err="failed to get container status \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": rpc error: code = NotFound desc = could not find container \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": container with ID starting with c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.393208 4758 scope.go:117] "RemoveContainer" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.393394 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} err="failed to get container status \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": rpc error: code = NotFound desc = could not find container \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": container with ID starting with 1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.393417 4758 scope.go:117] "RemoveContainer" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.394222 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} err="failed to get container status \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": rpc error: code = NotFound desc = could not find container \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": container with ID starting with 80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.394245 4758 scope.go:117] "RemoveContainer" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.394504 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} err="failed to get container status \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": rpc error: code = NotFound desc = could not find container \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": container with ID starting with 18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.394523 4758 scope.go:117] "RemoveContainer" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.394904 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} err="failed to get container status \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": rpc error: code = NotFound desc = could not find container \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": container with ID starting with ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.394989 4758 scope.go:117] "RemoveContainer" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396070 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} err="failed to get container status \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": rpc error: code = NotFound desc = could not find container \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": container with ID starting with 9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396118 4758 scope.go:117] "RemoveContainer" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396310 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} err="failed to get container status \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": rpc error: code = NotFound desc = could not find container \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": container with ID starting with 1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396329 4758 scope.go:117] "RemoveContainer" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396492 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} err="failed to get container status \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": rpc error: code = NotFound desc = could not find container \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": container with ID starting with a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396511 4758 scope.go:117] "RemoveContainer" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396691 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} err="failed to get container status \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": rpc error: code = NotFound desc = could not find container \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": container with ID starting with 2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396708 4758 scope.go:117] "RemoveContainer" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396885 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} err="failed to get container status \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": rpc error: code = NotFound desc = could not find container \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": container with ID starting with 9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.396903 4758 scope.go:117] "RemoveContainer" containerID="c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.397168 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b"} err="failed to get container status \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": rpc error: code = NotFound desc = could not find container \"c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b\": container with ID starting with c53ad30285c424adfb47b6f29b4e4f0a4dae165a62cdecf36ce107d52909413b not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.397215 4758 scope.go:117] "RemoveContainer" containerID="1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.397385 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3"} err="failed to get container status \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": rpc error: code = NotFound desc = could not find container \"1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3\": container with ID starting with 1902377dc21a02f820c4a85ce0369ccd7af75ddbaba5c978dfb69c8718dc75f3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.397407 4758 scope.go:117] "RemoveContainer" containerID="80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.397692 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f"} err="failed to get container status \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": rpc error: code = NotFound desc = could not find container \"80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f\": container with ID starting with 80575ffadc9f0481c3ecd415b79d1c36c5d251f85b17b41535d664af1b0a1a2f not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.397784 4758 scope.go:117] "RemoveContainer" containerID="18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398134 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d"} err="failed to get container status \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": rpc error: code = NotFound desc = could not find container \"18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d\": container with ID starting with 18ca5b2a07efb872ddf9dba598a90a69f934e1674991f100ef007158f90b224d not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398156 4758 scope.go:117] "RemoveContainer" containerID="ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398371 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252"} err="failed to get container status \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": rpc error: code = NotFound desc = could not find container \"ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252\": container with ID starting with ebe40bfeca562ede8870d9472125a2a57d33ea637483a74bae00f705a7ad0252 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398398 4758 scope.go:117] "RemoveContainer" containerID="9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398567 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69"} err="failed to get container status \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": rpc error: code = NotFound desc = could not find container \"9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69\": container with ID starting with 9032d8107809572bd8009a3b212a156201f1029a9384c79b57bb3b731cb77a69 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398591 4758 scope.go:117] "RemoveContainer" containerID="1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398744 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9"} err="failed to get container status \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": rpc error: code = NotFound desc = could not find container \"1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9\": container with ID starting with 1ecdc8adbd3eda14f07024430cb2e8a9ebcbd9a149f8766927d9f609f457e7a9 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398775 4758 scope.go:117] "RemoveContainer" containerID="a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.398988 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6"} err="failed to get container status \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": rpc error: code = NotFound desc = could not find container \"a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6\": container with ID starting with a3b9c6687d9e37bec6ec948e8ecafb458dd49711337b0e9ca2bf3cabfcb498b6 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.399007 4758 scope.go:117] "RemoveContainer" containerID="2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.399265 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3"} err="failed to get container status \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": rpc error: code = NotFound desc = could not find container \"2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3\": container with ID starting with 2a552e976207a30c2f24cf9e4db872df6f3be0a9440ce4eb536941107c7b2df3 not found: ID does not exist" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.399287 4758 scope.go:117] "RemoveContainer" containerID="9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62" Oct 04 11:00:03 crc kubenswrapper[4758]: I1004 11:00:03.399485 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62"} err="failed to get container status \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": rpc error: code = NotFound desc = could not find container \"9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62\": container with ID starting with 9adaf7cd5db1321bdd65cd2a0876f2b997f75957bb4a0c4822be4ad166be3e62 not found: ID does not exist" Oct 04 11:00:04 crc kubenswrapper[4758]: I1004 11:00:04.157180 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"0a5e5c2c60c0f231db98102eaee52a2ca652570b8e393132a90245e5a4f20179"} Oct 04 11:00:04 crc kubenswrapper[4758]: I1004 11:00:04.157226 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"4b948128d4064512fea854f0cc1b452c181d4399c631393c2d31151fcc388d6c"} Oct 04 11:00:04 crc kubenswrapper[4758]: I1004 11:00:04.157238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"627825ec4d82b1c5bff312260f489f230a479e1a00f140520ddebfbcf1c0ce66"} Oct 04 11:00:04 crc kubenswrapper[4758]: I1004 11:00:04.157250 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"13afb216285e4cb3b6365b33ab082edbd7b392da3a90118b1951bbd7dbf13dc8"} Oct 04 11:00:04 crc kubenswrapper[4758]: I1004 11:00:04.157260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"f74c5158dc07779585e583724a7d00a90b42ac2ae29b69a488858d2568c365c8"} Oct 04 11:00:04 crc kubenswrapper[4758]: I1004 11:00:04.157270 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"25bd45f6bc7aca6f4319848d492a6192af71978d28d2b1dccfb1dc0b49953d5a"} Oct 04 11:00:06 crc kubenswrapper[4758]: I1004 11:00:06.176211 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"8348de103611bfb00243e65d9bec1f4d7b8b5c2baf30200302cca96ba9d46865"} Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.197441 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" event={"ID":"29580f20-e100-43a2-a2ad-deb27a5e0340","Type":"ContainerStarted","Data":"a7da681896eed54339425891dcfa47237afa1831487e52a7977e55ca3f009e6b"} Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.197780 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.197796 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.197808 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.228387 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" podStartSLOduration=7.228369454 podStartE2EDuration="7.228369454s" podCreationTimestamp="2025-10-04 11:00:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:00:09.227588092 +0000 UTC m=+586.520239011" watchObservedRunningTime="2025-10-04 11:00:09.228369454 +0000 UTC m=+586.521020343" Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.235110 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:09 crc kubenswrapper[4758]: I1004 11:00:09.235722 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:18 crc kubenswrapper[4758]: I1004 11:00:18.325983 4758 scope.go:117] "RemoveContainer" containerID="1570a60d9f80108e4ebf690d60a5ecffd03622708015353cc0f5247388c750f8" Oct 04 11:00:18 crc kubenswrapper[4758]: E1004 11:00:18.326908 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bpvxr_openshift-multus(3875fe1d-de8c-4266-bf2b-e07c633b85dc)\"" pod="openshift-multus/multus-bpvxr" podUID="3875fe1d-de8c-4266-bf2b-e07c633b85dc" Oct 04 11:00:23 crc kubenswrapper[4758]: I1004 11:00:23.544460 4758 scope.go:117] "RemoveContainer" containerID="b80c58d308825f09f4b266c40d69ffd57667534dc74256a43337ff8029a8626e" Oct 04 11:00:24 crc kubenswrapper[4758]: I1004 11:00:24.293508 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/2.log" Oct 04 11:00:31 crc kubenswrapper[4758]: I1004 11:00:31.250351 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:00:31 crc kubenswrapper[4758]: I1004 11:00:31.251657 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:00:31 crc kubenswrapper[4758]: I1004 11:00:31.326716 4758 scope.go:117] "RemoveContainer" containerID="1570a60d9f80108e4ebf690d60a5ecffd03622708015353cc0f5247388c750f8" Oct 04 11:00:32 crc kubenswrapper[4758]: I1004 11:00:32.370284 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bpvxr_3875fe1d-de8c-4266-bf2b-e07c633b85dc/kube-multus/2.log" Oct 04 11:00:32 crc kubenswrapper[4758]: I1004 11:00:32.371041 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bpvxr" event={"ID":"3875fe1d-de8c-4266-bf2b-e07c633b85dc","Type":"ContainerStarted","Data":"d2421a13236caa4e0635f818d2606b311e2c18c6f7d150dcf346f2bb1efa469f"} Oct 04 11:00:32 crc kubenswrapper[4758]: I1004 11:00:32.826300 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mtnlv" Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.973781 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c"] Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.975272 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.978526 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.988135 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c"] Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.998060 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.998123 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z4hb\" (UniqueName: \"kubernetes.io/projected/b4165285-28ee-49e4-901c-06c614575ee9-kube-api-access-4z4hb\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:41 crc kubenswrapper[4758]: I1004 11:00:41.998295 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.099257 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.099354 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.099399 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z4hb\" (UniqueName: \"kubernetes.io/projected/b4165285-28ee-49e4-901c-06c614575ee9-kube-api-access-4z4hb\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.100149 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.100352 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.117436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z4hb\" (UniqueName: \"kubernetes.io/projected/b4165285-28ee-49e4-901c-06c614575ee9-kube-api-access-4z4hb\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.300330 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:42 crc kubenswrapper[4758]: I1004 11:00:42.756530 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c"] Oct 04 11:00:43 crc kubenswrapper[4758]: I1004 11:00:43.438342 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4165285-28ee-49e4-901c-06c614575ee9" containerID="4344e90282511c4a2e1cb1de457b1eea52e1f41eff75a3b9d5b5e88d293bfa0a" exitCode=0 Oct 04 11:00:43 crc kubenswrapper[4758]: I1004 11:00:43.438419 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" event={"ID":"b4165285-28ee-49e4-901c-06c614575ee9","Type":"ContainerDied","Data":"4344e90282511c4a2e1cb1de457b1eea52e1f41eff75a3b9d5b5e88d293bfa0a"} Oct 04 11:00:43 crc kubenswrapper[4758]: I1004 11:00:43.438500 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" event={"ID":"b4165285-28ee-49e4-901c-06c614575ee9","Type":"ContainerStarted","Data":"2bdeeb5a53880c69912f3bcf847d2b0af7feec8a78dd89f618b800440286a9fb"} Oct 04 11:00:45 crc kubenswrapper[4758]: I1004 11:00:45.452865 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4165285-28ee-49e4-901c-06c614575ee9" containerID="f53023eb5e006f4c3f12a2bf2ae48847690cbb5aab8229efc6d5bdec3ebc2f23" exitCode=0 Oct 04 11:00:45 crc kubenswrapper[4758]: I1004 11:00:45.452955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" event={"ID":"b4165285-28ee-49e4-901c-06c614575ee9","Type":"ContainerDied","Data":"f53023eb5e006f4c3f12a2bf2ae48847690cbb5aab8229efc6d5bdec3ebc2f23"} Oct 04 11:00:46 crc kubenswrapper[4758]: I1004 11:00:46.463267 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4165285-28ee-49e4-901c-06c614575ee9" containerID="99f9be75181c767d5fac0ab3673557e61ad5286ba205491bf23b129c79f93188" exitCode=0 Oct 04 11:00:46 crc kubenswrapper[4758]: I1004 11:00:46.463406 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" event={"ID":"b4165285-28ee-49e4-901c-06c614575ee9","Type":"ContainerDied","Data":"99f9be75181c767d5fac0ab3673557e61ad5286ba205491bf23b129c79f93188"} Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.737316 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.873938 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-bundle\") pod \"b4165285-28ee-49e4-901c-06c614575ee9\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.874000 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-util\") pod \"b4165285-28ee-49e4-901c-06c614575ee9\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.874050 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z4hb\" (UniqueName: \"kubernetes.io/projected/b4165285-28ee-49e4-901c-06c614575ee9-kube-api-access-4z4hb\") pod \"b4165285-28ee-49e4-901c-06c614575ee9\" (UID: \"b4165285-28ee-49e4-901c-06c614575ee9\") " Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.874979 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-bundle" (OuterVolumeSpecName: "bundle") pod "b4165285-28ee-49e4-901c-06c614575ee9" (UID: "b4165285-28ee-49e4-901c-06c614575ee9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.875403 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.881815 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4165285-28ee-49e4-901c-06c614575ee9-kube-api-access-4z4hb" (OuterVolumeSpecName: "kube-api-access-4z4hb") pod "b4165285-28ee-49e4-901c-06c614575ee9" (UID: "b4165285-28ee-49e4-901c-06c614575ee9"). InnerVolumeSpecName "kube-api-access-4z4hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.887308 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-util" (OuterVolumeSpecName: "util") pod "b4165285-28ee-49e4-901c-06c614575ee9" (UID: "b4165285-28ee-49e4-901c-06c614575ee9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.975962 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4165285-28ee-49e4-901c-06c614575ee9-util\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:47 crc kubenswrapper[4758]: I1004 11:00:47.976001 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z4hb\" (UniqueName: \"kubernetes.io/projected/b4165285-28ee-49e4-901c-06c614575ee9-kube-api-access-4z4hb\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:48 crc kubenswrapper[4758]: I1004 11:00:48.479180 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" event={"ID":"b4165285-28ee-49e4-901c-06c614575ee9","Type":"ContainerDied","Data":"2bdeeb5a53880c69912f3bcf847d2b0af7feec8a78dd89f618b800440286a9fb"} Oct 04 11:00:48 crc kubenswrapper[4758]: I1004 11:00:48.479240 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bdeeb5a53880c69912f3bcf847d2b0af7feec8a78dd89f618b800440286a9fb" Oct 04 11:00:48 crc kubenswrapper[4758]: I1004 11:00:48.479295 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.591462 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k"] Oct 04 11:00:51 crc kubenswrapper[4758]: E1004 11:00:51.591711 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="pull" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.591726 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="pull" Oct 04 11:00:51 crc kubenswrapper[4758]: E1004 11:00:51.591741 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="util" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.591749 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="util" Oct 04 11:00:51 crc kubenswrapper[4758]: E1004 11:00:51.591769 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="extract" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.591777 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="extract" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.591907 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4165285-28ee-49e4-901c-06c614575ee9" containerName="extract" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.592331 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.594697 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.595274 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.596284 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-gvhg5" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.615758 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k"] Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.721227 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qm4d\" (UniqueName: \"kubernetes.io/projected/e389e62d-20e1-448b-a6ce-a50609f8dfd0-kube-api-access-6qm4d\") pod \"nmstate-operator-858ddd8f98-4kt8k\" (UID: \"e389e62d-20e1-448b-a6ce-a50609f8dfd0\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.822086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qm4d\" (UniqueName: \"kubernetes.io/projected/e389e62d-20e1-448b-a6ce-a50609f8dfd0-kube-api-access-6qm4d\") pod \"nmstate-operator-858ddd8f98-4kt8k\" (UID: \"e389e62d-20e1-448b-a6ce-a50609f8dfd0\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.849775 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qm4d\" (UniqueName: \"kubernetes.io/projected/e389e62d-20e1-448b-a6ce-a50609f8dfd0-kube-api-access-6qm4d\") pod \"nmstate-operator-858ddd8f98-4kt8k\" (UID: \"e389e62d-20e1-448b-a6ce-a50609f8dfd0\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" Oct 04 11:00:51 crc kubenswrapper[4758]: I1004 11:00:51.910172 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" Oct 04 11:00:52 crc kubenswrapper[4758]: I1004 11:00:52.177714 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k"] Oct 04 11:00:52 crc kubenswrapper[4758]: W1004 11:00:52.182666 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode389e62d_20e1_448b_a6ce_a50609f8dfd0.slice/crio-3589c9f24db388b7551ecc16dcda67223bf21ba485760ede352bc9e93418008b WatchSource:0}: Error finding container 3589c9f24db388b7551ecc16dcda67223bf21ba485760ede352bc9e93418008b: Status 404 returned error can't find the container with id 3589c9f24db388b7551ecc16dcda67223bf21ba485760ede352bc9e93418008b Oct 04 11:00:52 crc kubenswrapper[4758]: I1004 11:00:52.500156 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" event={"ID":"e389e62d-20e1-448b-a6ce-a50609f8dfd0","Type":"ContainerStarted","Data":"3589c9f24db388b7551ecc16dcda67223bf21ba485760ede352bc9e93418008b"} Oct 04 11:00:59 crc kubenswrapper[4758]: I1004 11:00:59.546075 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" event={"ID":"e389e62d-20e1-448b-a6ce-a50609f8dfd0","Type":"ContainerStarted","Data":"56712805dfc89869c73a9146132fbac8ceeedc5997364de3ce70b6412c343419"} Oct 04 11:00:59 crc kubenswrapper[4758]: I1004 11:00:59.567493 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-4kt8k" podStartSLOduration=1.4405905030000001 podStartE2EDuration="8.567475206s" podCreationTimestamp="2025-10-04 11:00:51 +0000 UTC" firstStartedPulling="2025-10-04 11:00:52.184562897 +0000 UTC m=+629.477213786" lastFinishedPulling="2025-10-04 11:00:59.31144756 +0000 UTC m=+636.604098489" observedRunningTime="2025-10-04 11:00:59.566372865 +0000 UTC m=+636.859023814" watchObservedRunningTime="2025-10-04 11:00:59.567475206 +0000 UTC m=+636.860126115" Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.249258 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.249337 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.249386 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.249921 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4a4ba30925ab5365601b714b606e8b4bf9695f793e826fa68e869bdae0e9d640"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.249972 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://4a4ba30925ab5365601b714b606e8b4bf9695f793e826fa68e869bdae0e9d640" gracePeriod=600 Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.558348 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="4a4ba30925ab5365601b714b606e8b4bf9695f793e826fa68e869bdae0e9d640" exitCode=0 Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.558404 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"4a4ba30925ab5365601b714b606e8b4bf9695f793e826fa68e869bdae0e9d640"} Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.558694 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"5c234c874dc874e8dd12e5faf6f9512499bc90758f95287be1214207762ac29f"} Oct 04 11:01:01 crc kubenswrapper[4758]: I1004 11:01:01.558733 4758 scope.go:117] "RemoveContainer" containerID="770a65d2316af3d496832d48f9bf7d288866132c38f07143143f0ef4b5eb238f" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.583894 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.585380 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.587893 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-hhrt6" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.595119 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.614193 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vx78k"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.614790 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.628362 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.629184 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.638234 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651305 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-dbus-socket\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651359 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ce3569b2-1a2d-4d82-8856-69738a2eb8dd-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-lzss8\" (UID: \"ce3569b2-1a2d-4d82-8856-69738a2eb8dd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651509 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgd8d\" (UniqueName: \"kubernetes.io/projected/ce3569b2-1a2d-4d82-8856-69738a2eb8dd-kube-api-access-fgd8d\") pod \"nmstate-webhook-6cdbc54649-lzss8\" (UID: \"ce3569b2-1a2d-4d82-8856-69738a2eb8dd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651552 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkxk7\" (UniqueName: \"kubernetes.io/projected/ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24-kube-api-access-jkxk7\") pod \"nmstate-metrics-fdff9cb8d-f6nvf\" (UID: \"ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-nmstate-lock\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651606 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-ovs-socket\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.651628 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl6h7\" (UniqueName: \"kubernetes.io/projected/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-kube-api-access-xl6h7\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.666547 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752336 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-dbus-socket\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ce3569b2-1a2d-4d82-8856-69738a2eb8dd-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-lzss8\" (UID: \"ce3569b2-1a2d-4d82-8856-69738a2eb8dd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752461 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgd8d\" (UniqueName: \"kubernetes.io/projected/ce3569b2-1a2d-4d82-8856-69738a2eb8dd-kube-api-access-fgd8d\") pod \"nmstate-webhook-6cdbc54649-lzss8\" (UID: \"ce3569b2-1a2d-4d82-8856-69738a2eb8dd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752484 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkxk7\" (UniqueName: \"kubernetes.io/projected/ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24-kube-api-access-jkxk7\") pod \"nmstate-metrics-fdff9cb8d-f6nvf\" (UID: \"ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752504 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-nmstate-lock\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752532 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-ovs-socket\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752556 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl6h7\" (UniqueName: \"kubernetes.io/projected/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-kube-api-access-xl6h7\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752733 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-dbus-socket\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752789 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-ovs-socket\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.752794 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-nmstate-lock\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.753694 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.754291 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.761488 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cp8pj" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.761570 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.761771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ce3569b2-1a2d-4d82-8856-69738a2eb8dd-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-lzss8\" (UID: \"ce3569b2-1a2d-4d82-8856-69738a2eb8dd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.766286 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.778253 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz"] Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.785573 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl6h7\" (UniqueName: \"kubernetes.io/projected/30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb-kube-api-access-xl6h7\") pod \"nmstate-handler-vx78k\" (UID: \"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb\") " pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.786152 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgd8d\" (UniqueName: \"kubernetes.io/projected/ce3569b2-1a2d-4d82-8856-69738a2eb8dd-kube-api-access-fgd8d\") pod \"nmstate-webhook-6cdbc54649-lzss8\" (UID: \"ce3569b2-1a2d-4d82-8856-69738a2eb8dd\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.792835 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkxk7\" (UniqueName: \"kubernetes.io/projected/ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24-kube-api-access-jkxk7\") pod \"nmstate-metrics-fdff9cb8d-f6nvf\" (UID: \"ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.876284 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fb02372-f4a6-408a-8c69-ddee8717758e-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.876342 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb02372-f4a6-408a-8c69-ddee8717758e-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.876378 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55gzb\" (UniqueName: \"kubernetes.io/projected/5fb02372-f4a6-408a-8c69-ddee8717758e-kube-api-access-55gzb\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.901507 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.932198 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.942038 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.977231 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55gzb\" (UniqueName: \"kubernetes.io/projected/5fb02372-f4a6-408a-8c69-ddee8717758e-kube-api-access-55gzb\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.977329 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fb02372-f4a6-408a-8c69-ddee8717758e-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.977375 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb02372-f4a6-408a-8c69-ddee8717758e-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.978548 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fb02372-f4a6-408a-8c69-ddee8717758e-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:02 crc kubenswrapper[4758]: I1004 11:01:02.983599 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb02372-f4a6-408a-8c69-ddee8717758e-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.004704 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55gzb\" (UniqueName: \"kubernetes.io/projected/5fb02372-f4a6-408a-8c69-ddee8717758e-kube-api-access-55gzb\") pod \"nmstate-console-plugin-6b874cbd85-76lpz\" (UID: \"5fb02372-f4a6-408a-8c69-ddee8717758e\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.024251 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5d9bf7d847-rh8nc"] Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.025592 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.048811 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d9bf7d847-rh8nc"] Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078468 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/697cca45-a81d-40cd-8fbe-c877c73569c9-console-oauth-config\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078853 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/697cca45-a81d-40cd-8fbe-c877c73569c9-console-serving-cert\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078881 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d82j\" (UniqueName: \"kubernetes.io/projected/697cca45-a81d-40cd-8fbe-c877c73569c9-kube-api-access-5d82j\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078902 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-trusted-ca-bundle\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078921 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-console-config\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078957 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-service-ca\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.078997 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-oauth-serving-cert\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.086035 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179586 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/697cca45-a81d-40cd-8fbe-c877c73569c9-console-serving-cert\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179616 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/697cca45-a81d-40cd-8fbe-c877c73569c9-console-oauth-config\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d82j\" (UniqueName: \"kubernetes.io/projected/697cca45-a81d-40cd-8fbe-c877c73569c9-kube-api-access-5d82j\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179660 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-trusted-ca-bundle\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179676 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-console-config\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179705 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-service-ca\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.179737 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-oauth-serving-cert\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.180621 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-oauth-serving-cert\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.181163 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-console-config\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.181268 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-trusted-ca-bundle\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.181707 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/697cca45-a81d-40cd-8fbe-c877c73569c9-service-ca\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.194311 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/697cca45-a81d-40cd-8fbe-c877c73569c9-console-oauth-config\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.198483 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/697cca45-a81d-40cd-8fbe-c877c73569c9-console-serving-cert\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.199811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d82j\" (UniqueName: \"kubernetes.io/projected/697cca45-a81d-40cd-8fbe-c877c73569c9-kube-api-access-5d82j\") pod \"console-5d9bf7d847-rh8nc\" (UID: \"697cca45-a81d-40cd-8fbe-c877c73569c9\") " pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.214030 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf"] Oct 04 11:01:03 crc kubenswrapper[4758]: W1004 11:01:03.223117 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad5892a8_3bbe_43ab_9544_5ca0f5bcbf24.slice/crio-1eb6e13d41b4b022e0c278b48e8da0e56329905ea2861695a0a69374af5275ad WatchSource:0}: Error finding container 1eb6e13d41b4b022e0c278b48e8da0e56329905ea2861695a0a69374af5275ad: Status 404 returned error can't find the container with id 1eb6e13d41b4b022e0c278b48e8da0e56329905ea2861695a0a69374af5275ad Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.344546 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.495839 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8"] Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.540259 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz"] Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.548197 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d9bf7d847-rh8nc"] Oct 04 11:01:03 crc kubenswrapper[4758]: W1004 11:01:03.552166 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod697cca45_a81d_40cd_8fbe_c877c73569c9.slice/crio-dec216af86bf136a1984dfbfe30f806af867cdbb6a166f34a10d8bb2df47c6b1 WatchSource:0}: Error finding container dec216af86bf136a1984dfbfe30f806af867cdbb6a166f34a10d8bb2df47c6b1: Status 404 returned error can't find the container with id dec216af86bf136a1984dfbfe30f806af867cdbb6a166f34a10d8bb2df47c6b1 Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.575039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" event={"ID":"ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24","Type":"ContainerStarted","Data":"1eb6e13d41b4b022e0c278b48e8da0e56329905ea2861695a0a69374af5275ad"} Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.576045 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d9bf7d847-rh8nc" event={"ID":"697cca45-a81d-40cd-8fbe-c877c73569c9","Type":"ContainerStarted","Data":"dec216af86bf136a1984dfbfe30f806af867cdbb6a166f34a10d8bb2df47c6b1"} Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.578367 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" event={"ID":"5fb02372-f4a6-408a-8c69-ddee8717758e","Type":"ContainerStarted","Data":"92d95161319f031c29effe0f0b63dc8f2965771f8358ea3572777f95d6e5980f"} Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.579055 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vx78k" event={"ID":"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb","Type":"ContainerStarted","Data":"139d8978aaf9e6c1cc432a256306e00a8606c62054c4b48f8e9a38a864fb964c"} Oct 04 11:01:03 crc kubenswrapper[4758]: I1004 11:01:03.580640 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" event={"ID":"ce3569b2-1a2d-4d82-8856-69738a2eb8dd","Type":"ContainerStarted","Data":"251cc4c0f8b2bd78e2139e5e5e908e1d9c6f39106b52e86fb39e51027b9a3904"} Oct 04 11:01:04 crc kubenswrapper[4758]: I1004 11:01:04.589083 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d9bf7d847-rh8nc" event={"ID":"697cca45-a81d-40cd-8fbe-c877c73569c9","Type":"ContainerStarted","Data":"9213d1c3a12cc95682e93e71920c0a8d660e6d180b9c99b1eb85e5d5b1738160"} Oct 04 11:01:04 crc kubenswrapper[4758]: I1004 11:01:04.613140 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5d9bf7d847-rh8nc" podStartSLOduration=1.6130853680000001 podStartE2EDuration="1.613085368s" podCreationTimestamp="2025-10-04 11:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:01:04.612564654 +0000 UTC m=+641.905215583" watchObservedRunningTime="2025-10-04 11:01:04.613085368 +0000 UTC m=+641.905736297" Oct 04 11:01:08 crc kubenswrapper[4758]: I1004 11:01:08.616036 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vx78k" event={"ID":"30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb","Type":"ContainerStarted","Data":"45441597befec171e15f0746f57c156c58d60b2d8540e7fddcd19dfb718f43cd"} Oct 04 11:01:08 crc kubenswrapper[4758]: I1004 11:01:08.616452 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:08 crc kubenswrapper[4758]: I1004 11:01:08.617674 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" event={"ID":"ce3569b2-1a2d-4d82-8856-69738a2eb8dd","Type":"ContainerStarted","Data":"36eafb3b51243741e6cced675df47041a232f0824bfb2756b0a802ea85c47196"} Oct 04 11:01:08 crc kubenswrapper[4758]: I1004 11:01:08.618648 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:08 crc kubenswrapper[4758]: I1004 11:01:08.638200 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vx78k" podStartSLOduration=2.239857252 podStartE2EDuration="6.638179347s" podCreationTimestamp="2025-10-04 11:01:02 +0000 UTC" firstStartedPulling="2025-10-04 11:01:03.005199255 +0000 UTC m=+640.297850144" lastFinishedPulling="2025-10-04 11:01:07.40352135 +0000 UTC m=+644.696172239" observedRunningTime="2025-10-04 11:01:08.629194035 +0000 UTC m=+645.921844944" watchObservedRunningTime="2025-10-04 11:01:08.638179347 +0000 UTC m=+645.930830246" Oct 04 11:01:08 crc kubenswrapper[4758]: I1004 11:01:08.645178 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" podStartSLOduration=2.753111287 podStartE2EDuration="6.645144582s" podCreationTimestamp="2025-10-04 11:01:02 +0000 UTC" firstStartedPulling="2025-10-04 11:01:03.519496549 +0000 UTC m=+640.812147438" lastFinishedPulling="2025-10-04 11:01:07.411529834 +0000 UTC m=+644.704180733" observedRunningTime="2025-10-04 11:01:08.643064983 +0000 UTC m=+645.935715902" watchObservedRunningTime="2025-10-04 11:01:08.645144582 +0000 UTC m=+645.937795511" Oct 04 11:01:09 crc kubenswrapper[4758]: I1004 11:01:09.628571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" event={"ID":"5fb02372-f4a6-408a-8c69-ddee8717758e","Type":"ContainerStarted","Data":"85e96f45f3f34561afb79d04fcaac2627affc80e8bbb0dfe9d7e73a2d6d1d547"} Oct 04 11:01:09 crc kubenswrapper[4758]: I1004 11:01:09.655855 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-76lpz" podStartSLOduration=2.496739521 podStartE2EDuration="7.6558365s" podCreationTimestamp="2025-10-04 11:01:02 +0000 UTC" firstStartedPulling="2025-10-04 11:01:03.542719069 +0000 UTC m=+640.835369968" lastFinishedPulling="2025-10-04 11:01:08.701816058 +0000 UTC m=+645.994466947" observedRunningTime="2025-10-04 11:01:09.653506975 +0000 UTC m=+646.946157874" watchObservedRunningTime="2025-10-04 11:01:09.6558365 +0000 UTC m=+646.948487389" Oct 04 11:01:12 crc kubenswrapper[4758]: I1004 11:01:12.965548 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vx78k" Oct 04 11:01:13 crc kubenswrapper[4758]: I1004 11:01:13.345279 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:13 crc kubenswrapper[4758]: I1004 11:01:13.345343 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:13 crc kubenswrapper[4758]: I1004 11:01:13.355613 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:13 crc kubenswrapper[4758]: I1004 11:01:13.662638 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5d9bf7d847-rh8nc" Oct 04 11:01:13 crc kubenswrapper[4758]: I1004 11:01:13.737075 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pqnhz"] Oct 04 11:01:15 crc kubenswrapper[4758]: I1004 11:01:15.667987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" event={"ID":"ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24","Type":"ContainerStarted","Data":"5d8dd28e9056f8a2b94bd36a003cab4ca9c884104c8d228fba7071443f149a87"} Oct 04 11:01:19 crc kubenswrapper[4758]: I1004 11:01:19.693822 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" event={"ID":"ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24","Type":"ContainerStarted","Data":"ce1741852b0c228bb23f206073f4dac097554d2415e8c29d4b17cd8a07b85c65"} Oct 04 11:01:22 crc kubenswrapper[4758]: I1004 11:01:22.954995 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-lzss8" Oct 04 11:01:22 crc kubenswrapper[4758]: I1004 11:01:22.988379 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-f6nvf" podStartSLOduration=5.136005381 podStartE2EDuration="20.988352045s" podCreationTimestamp="2025-10-04 11:01:02 +0000 UTC" firstStartedPulling="2025-10-04 11:01:03.225307625 +0000 UTC m=+640.517958514" lastFinishedPulling="2025-10-04 11:01:19.077654289 +0000 UTC m=+656.370305178" observedRunningTime="2025-10-04 11:01:19.720543101 +0000 UTC m=+657.013193990" watchObservedRunningTime="2025-10-04 11:01:22.988352045 +0000 UTC m=+660.281002964" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.165184 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk"] Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.166752 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.168487 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.173760 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk"] Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.258772 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.258813 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.258844 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzmd2\" (UniqueName: \"kubernetes.io/projected/54cccc46-3ea9-493b-b2fe-20e90218f026-kube-api-access-zzmd2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.359538 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.359771 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.359800 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzmd2\" (UniqueName: \"kubernetes.io/projected/54cccc46-3ea9-493b-b2fe-20e90218f026-kube-api-access-zzmd2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.360140 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.360415 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.383753 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzmd2\" (UniqueName: \"kubernetes.io/projected/54cccc46-3ea9-493b-b2fe-20e90218f026-kube-api-access-zzmd2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.537558 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:36 crc kubenswrapper[4758]: I1004 11:01:36.926740 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk"] Oct 04 11:01:36 crc kubenswrapper[4758]: W1004 11:01:36.932213 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54cccc46_3ea9_493b_b2fe_20e90218f026.slice/crio-4410b195ddaa5d707a771a40e2e2012290de9e1ef503223303813e10a0ba70f4 WatchSource:0}: Error finding container 4410b195ddaa5d707a771a40e2e2012290de9e1ef503223303813e10a0ba70f4: Status 404 returned error can't find the container with id 4410b195ddaa5d707a771a40e2e2012290de9e1ef503223303813e10a0ba70f4 Oct 04 11:01:37 crc kubenswrapper[4758]: I1004 11:01:37.824731 4758 generic.go:334] "Generic (PLEG): container finished" podID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerID="1df281f8627f09e348007cdcf1dd885bf5a2f25c4c21b48a8429030d5f951164" exitCode=0 Oct 04 11:01:37 crc kubenswrapper[4758]: I1004 11:01:37.824837 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" event={"ID":"54cccc46-3ea9-493b-b2fe-20e90218f026","Type":"ContainerDied","Data":"1df281f8627f09e348007cdcf1dd885bf5a2f25c4c21b48a8429030d5f951164"} Oct 04 11:01:37 crc kubenswrapper[4758]: I1004 11:01:37.827239 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" event={"ID":"54cccc46-3ea9-493b-b2fe-20e90218f026","Type":"ContainerStarted","Data":"4410b195ddaa5d707a771a40e2e2012290de9e1ef503223303813e10a0ba70f4"} Oct 04 11:01:38 crc kubenswrapper[4758]: I1004 11:01:38.804506 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-pqnhz" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerName="console" containerID="cri-o://8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6" gracePeriod=15 Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.204896 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pqnhz_aed3108c-fee1-4f60-b31f-ba579cbb34b6/console/0.log" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.205229 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343587 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-oauth-serving-cert\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343692 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-config\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343729 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-serving-cert\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343753 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-service-ca\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343790 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7mhc\" (UniqueName: \"kubernetes.io/projected/aed3108c-fee1-4f60-b31f-ba579cbb34b6-kube-api-access-n7mhc\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343827 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-trusted-ca-bundle\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.343868 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-oauth-config\") pod \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\" (UID: \"aed3108c-fee1-4f60-b31f-ba579cbb34b6\") " Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.344318 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.344615 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-service-ca" (OuterVolumeSpecName: "service-ca") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.344844 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-config" (OuterVolumeSpecName: "console-config") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.345056 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.350682 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.351167 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed3108c-fee1-4f60-b31f-ba579cbb34b6-kube-api-access-n7mhc" (OuterVolumeSpecName: "kube-api-access-n7mhc") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "kube-api-access-n7mhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.351592 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "aed3108c-fee1-4f60-b31f-ba579cbb34b6" (UID: "aed3108c-fee1-4f60-b31f-ba579cbb34b6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445395 4758 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445437 4758 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445453 4758 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445465 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7mhc\" (UniqueName: \"kubernetes.io/projected/aed3108c-fee1-4f60-b31f-ba579cbb34b6-kube-api-access-n7mhc\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445479 4758 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445491 4758 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/aed3108c-fee1-4f60-b31f-ba579cbb34b6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.445503 4758 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/aed3108c-fee1-4f60-b31f-ba579cbb34b6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.840031 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-pqnhz_aed3108c-fee1-4f60-b31f-ba579cbb34b6/console/0.log" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.840220 4758 generic.go:334] "Generic (PLEG): container finished" podID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerID="8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6" exitCode=2 Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.840453 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-pqnhz" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.840451 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pqnhz" event={"ID":"aed3108c-fee1-4f60-b31f-ba579cbb34b6","Type":"ContainerDied","Data":"8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6"} Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.840694 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-pqnhz" event={"ID":"aed3108c-fee1-4f60-b31f-ba579cbb34b6","Type":"ContainerDied","Data":"f061625caf05c1e25ab4e04180da4e4ab9cf9c5bd9ed7fb1d57daf01aeb0c964"} Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.840724 4758 scope.go:117] "RemoveContainer" containerID="8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.844505 4758 generic.go:334] "Generic (PLEG): container finished" podID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerID="d4f9b3e6b6252e68cf9bddd6b18fc3df772d0c0d30629f11bb67a3afacfc4114" exitCode=0 Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.844541 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" event={"ID":"54cccc46-3ea9-493b-b2fe-20e90218f026","Type":"ContainerDied","Data":"d4f9b3e6b6252e68cf9bddd6b18fc3df772d0c0d30629f11bb67a3afacfc4114"} Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.871144 4758 scope.go:117] "RemoveContainer" containerID="8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6" Oct 04 11:01:39 crc kubenswrapper[4758]: E1004 11:01:39.874674 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6\": container with ID starting with 8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6 not found: ID does not exist" containerID="8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.874729 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6"} err="failed to get container status \"8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6\": rpc error: code = NotFound desc = could not find container \"8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6\": container with ID starting with 8614bca33cec495db05c1b092f2f267b00f4e08fa139994e9bf4fb12497c64b6 not found: ID does not exist" Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.889602 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-pqnhz"] Oct 04 11:01:39 crc kubenswrapper[4758]: I1004 11:01:39.895539 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-pqnhz"] Oct 04 11:01:40 crc kubenswrapper[4758]: I1004 11:01:40.851715 4758 generic.go:334] "Generic (PLEG): container finished" podID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerID="a46c2761826141ea2ea422bed1e24e7018042e4920fa389e6bdaec4b09d938e2" exitCode=0 Oct 04 11:01:40 crc kubenswrapper[4758]: I1004 11:01:40.851863 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" event={"ID":"54cccc46-3ea9-493b-b2fe-20e90218f026","Type":"ContainerDied","Data":"a46c2761826141ea2ea422bed1e24e7018042e4920fa389e6bdaec4b09d938e2"} Oct 04 11:01:41 crc kubenswrapper[4758]: I1004 11:01:41.339315 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" path="/var/lib/kubelet/pods/aed3108c-fee1-4f60-b31f-ba579cbb34b6/volumes" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.120340 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.280904 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-util\") pod \"54cccc46-3ea9-493b-b2fe-20e90218f026\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.280983 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzmd2\" (UniqueName: \"kubernetes.io/projected/54cccc46-3ea9-493b-b2fe-20e90218f026-kube-api-access-zzmd2\") pod \"54cccc46-3ea9-493b-b2fe-20e90218f026\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.281063 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-bundle\") pod \"54cccc46-3ea9-493b-b2fe-20e90218f026\" (UID: \"54cccc46-3ea9-493b-b2fe-20e90218f026\") " Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.283178 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-bundle" (OuterVolumeSpecName: "bundle") pod "54cccc46-3ea9-493b-b2fe-20e90218f026" (UID: "54cccc46-3ea9-493b-b2fe-20e90218f026"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.290178 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cccc46-3ea9-493b-b2fe-20e90218f026-kube-api-access-zzmd2" (OuterVolumeSpecName: "kube-api-access-zzmd2") pod "54cccc46-3ea9-493b-b2fe-20e90218f026" (UID: "54cccc46-3ea9-493b-b2fe-20e90218f026"). InnerVolumeSpecName "kube-api-access-zzmd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.382484 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzmd2\" (UniqueName: \"kubernetes.io/projected/54cccc46-3ea9-493b-b2fe-20e90218f026-kube-api-access-zzmd2\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.382541 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.733065 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-util" (OuterVolumeSpecName: "util") pod "54cccc46-3ea9-493b-b2fe-20e90218f026" (UID: "54cccc46-3ea9-493b-b2fe-20e90218f026"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.788017 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54cccc46-3ea9-493b-b2fe-20e90218f026-util\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.870187 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" event={"ID":"54cccc46-3ea9-493b-b2fe-20e90218f026","Type":"ContainerDied","Data":"4410b195ddaa5d707a771a40e2e2012290de9e1ef503223303813e10a0ba70f4"} Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.870234 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4410b195ddaa5d707a771a40e2e2012290de9e1ef503223303813e10a0ba70f4" Oct 04 11:01:42 crc kubenswrapper[4758]: I1004 11:01:42.870342 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.222748 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq"] Oct 04 11:01:51 crc kubenswrapper[4758]: E1004 11:01:51.223481 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="extract" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223492 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="extract" Oct 04 11:01:51 crc kubenswrapper[4758]: E1004 11:01:51.223500 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="pull" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223506 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="pull" Oct 04 11:01:51 crc kubenswrapper[4758]: E1004 11:01:51.223519 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerName="console" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223524 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerName="console" Oct 04 11:01:51 crc kubenswrapper[4758]: E1004 11:01:51.223542 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="util" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223548 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="util" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223634 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed3108c-fee1-4f60-b31f-ba579cbb34b6" containerName="console" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223645 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cccc46-3ea9-493b-b2fe-20e90218f026" containerName="extract" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.223978 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.229781 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-bc9bh" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.230482 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.231366 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.232285 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.233583 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.245981 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq"] Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.391756 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fa4aec47-b441-4630-91a5-f092db76a1c5-apiservice-cert\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.391860 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlkqn\" (UniqueName: \"kubernetes.io/projected/fa4aec47-b441-4630-91a5-f092db76a1c5-kube-api-access-rlkqn\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.391988 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fa4aec47-b441-4630-91a5-f092db76a1c5-webhook-cert\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.492989 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fa4aec47-b441-4630-91a5-f092db76a1c5-webhook-cert\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.493043 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fa4aec47-b441-4630-91a5-f092db76a1c5-apiservice-cert\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.493117 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlkqn\" (UniqueName: \"kubernetes.io/projected/fa4aec47-b441-4630-91a5-f092db76a1c5-kube-api-access-rlkqn\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.498652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fa4aec47-b441-4630-91a5-f092db76a1c5-webhook-cert\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.507683 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fa4aec47-b441-4630-91a5-f092db76a1c5-apiservice-cert\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.513657 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlkqn\" (UniqueName: \"kubernetes.io/projected/fa4aec47-b441-4630-91a5-f092db76a1c5-kube-api-access-rlkqn\") pod \"metallb-operator-controller-manager-5d8b4d9f4c-wngnq\" (UID: \"fa4aec47-b441-4630-91a5-f092db76a1c5\") " pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.541620 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.730375 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97"] Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.731829 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.739909 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.742068 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-h6hxz" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.747837 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.758520 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97"] Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.898581 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-apiservice-cert\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.898619 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbkt8\" (UniqueName: \"kubernetes.io/projected/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-kube-api-access-kbkt8\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.898639 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-webhook-cert\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.999432 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbkt8\" (UniqueName: \"kubernetes.io/projected/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-kube-api-access-kbkt8\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.999472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-webhook-cert\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:51 crc kubenswrapper[4758]: I1004 11:01:51.999545 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-apiservice-cert\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.000287 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq"] Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.005138 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-apiservice-cert\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.005396 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-webhook-cert\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.026466 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbkt8\" (UniqueName: \"kubernetes.io/projected/fe3ea4e8-b319-41f2-902b-e3b5dcde02d3-kube-api-access-kbkt8\") pod \"metallb-operator-webhook-server-6b56c7d7-fmn97\" (UID: \"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3\") " pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.045334 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.271788 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97"] Oct 04 11:01:52 crc kubenswrapper[4758]: W1004 11:01:52.278993 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe3ea4e8_b319_41f2_902b_e3b5dcde02d3.slice/crio-b25ac4632ee645aea31b5898d0808f16be7f628938ba35cec5fc4d6ca2060a5c WatchSource:0}: Error finding container b25ac4632ee645aea31b5898d0808f16be7f628938ba35cec5fc4d6ca2060a5c: Status 404 returned error can't find the container with id b25ac4632ee645aea31b5898d0808f16be7f628938ba35cec5fc4d6ca2060a5c Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.920787 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" event={"ID":"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3","Type":"ContainerStarted","Data":"b25ac4632ee645aea31b5898d0808f16be7f628938ba35cec5fc4d6ca2060a5c"} Oct 04 11:01:52 crc kubenswrapper[4758]: I1004 11:01:52.922319 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" event={"ID":"fa4aec47-b441-4630-91a5-f092db76a1c5","Type":"ContainerStarted","Data":"0477756bda6a2e00e884004053fcf1b7fa07044b45565a0d23f65135fd92f604"} Oct 04 11:01:57 crc kubenswrapper[4758]: I1004 11:01:57.954504 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" event={"ID":"fe3ea4e8-b319-41f2-902b-e3b5dcde02d3","Type":"ContainerStarted","Data":"3b78ee9e2fd40d1d9f8d03fbb1952167b53df941ac4077fce1498181f9d88c8d"} Oct 04 11:01:57 crc kubenswrapper[4758]: I1004 11:01:57.954941 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:01:57 crc kubenswrapper[4758]: I1004 11:01:57.956811 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" event={"ID":"fa4aec47-b441-4630-91a5-f092db76a1c5","Type":"ContainerStarted","Data":"89f87b04a6b6c72fd302f5ed79c30bf9edab49b18a54ce15bc8ae7a67e655822"} Oct 04 11:01:57 crc kubenswrapper[4758]: I1004 11:01:57.956924 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:01:57 crc kubenswrapper[4758]: I1004 11:01:57.969626 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" podStartSLOduration=2.170044843 podStartE2EDuration="6.969613243s" podCreationTimestamp="2025-10-04 11:01:51 +0000 UTC" firstStartedPulling="2025-10-04 11:01:52.281393973 +0000 UTC m=+689.574044862" lastFinishedPulling="2025-10-04 11:01:57.080962373 +0000 UTC m=+694.373613262" observedRunningTime="2025-10-04 11:01:57.969357796 +0000 UTC m=+695.262008685" watchObservedRunningTime="2025-10-04 11:01:57.969613243 +0000 UTC m=+695.262264132" Oct 04 11:01:57 crc kubenswrapper[4758]: I1004 11:01:57.992268 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" podStartSLOduration=1.943478497 podStartE2EDuration="6.992246614s" podCreationTimestamp="2025-10-04 11:01:51 +0000 UTC" firstStartedPulling="2025-10-04 11:01:52.013161468 +0000 UTC m=+689.305812357" lastFinishedPulling="2025-10-04 11:01:57.061929585 +0000 UTC m=+694.354580474" observedRunningTime="2025-10-04 11:01:57.989027012 +0000 UTC m=+695.281677911" watchObservedRunningTime="2025-10-04 11:01:57.992246614 +0000 UTC m=+695.284897503" Oct 04 11:02:12 crc kubenswrapper[4758]: I1004 11:02:12.056465 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6b56c7d7-fmn97" Oct 04 11:02:31 crc kubenswrapper[4758]: I1004 11:02:31.545003 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d8b4d9f4c-wngnq" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.275480 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-ssxwx"] Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.278017 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.280614 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.280778 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.280907 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6jc2h" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.284702 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt"] Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.285938 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.287174 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.298031 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt"] Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408618 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-sockets\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408692 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg7td\" (UniqueName: \"kubernetes.io/projected/a20453a0-6ddc-44e5-bd0a-9615f3607127-kube-api-access-zg7td\") pod \"frr-k8s-webhook-server-64bf5d555-48dkt\" (UID: \"a20453a0-6ddc-44e5-bd0a-9615f3607127\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408722 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-metrics-certs\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408747 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-conf\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408778 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwqxj\" (UniqueName: \"kubernetes.io/projected/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-kube-api-access-zwqxj\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408803 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-metrics\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408823 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-reloader\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408845 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-startup\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.408893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a20453a0-6ddc-44e5-bd0a-9615f3607127-cert\") pod \"frr-k8s-webhook-server-64bf5d555-48dkt\" (UID: \"a20453a0-6ddc-44e5-bd0a-9615f3607127\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.410730 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bz5tp"] Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.411535 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.413697 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.413844 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.413967 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-q827h" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.414940 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.417169 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-cnb79"] Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.418035 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.419067 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.437247 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cnb79"] Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510128 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a20453a0-6ddc-44e5-bd0a-9615f3607127-cert\") pod \"frr-k8s-webhook-server-64bf5d555-48dkt\" (UID: \"a20453a0-6ddc-44e5-bd0a-9615f3607127\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510172 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510203 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72277c69-780e-4dcf-a26b-e49001cbf6d1-metallb-excludel2\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510221 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-sockets\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510241 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/063f91cf-fe76-4bc6-a070-ccdf43c43f02-metrics-certs\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510266 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg7td\" (UniqueName: \"kubernetes.io/projected/a20453a0-6ddc-44e5-bd0a-9615f3607127-kube-api-access-zg7td\") pod \"frr-k8s-webhook-server-64bf5d555-48dkt\" (UID: \"a20453a0-6ddc-44e5-bd0a-9615f3607127\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510280 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-metrics-certs\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510296 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-metrics-certs\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510311 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-conf\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5px75\" (UniqueName: \"kubernetes.io/projected/063f91cf-fe76-4bc6-a070-ccdf43c43f02-kube-api-access-5px75\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510352 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwqxj\" (UniqueName: \"kubernetes.io/projected/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-kube-api-access-zwqxj\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510368 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-metrics\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510387 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-reloader\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510403 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-startup\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510429 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvw4\" (UniqueName: \"kubernetes.io/projected/72277c69-780e-4dcf-a26b-e49001cbf6d1-kube-api-access-pwvw4\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.510446 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/063f91cf-fe76-4bc6-a070-ccdf43c43f02-cert\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.511642 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-sockets\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.511665 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-metrics\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.511833 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-conf\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.511943 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-reloader\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.518768 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-frr-startup\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.533640 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-metrics-certs\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.534700 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a20453a0-6ddc-44e5-bd0a-9615f3607127-cert\") pod \"frr-k8s-webhook-server-64bf5d555-48dkt\" (UID: \"a20453a0-6ddc-44e5-bd0a-9615f3607127\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.538548 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwqxj\" (UniqueName: \"kubernetes.io/projected/6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e-kube-api-access-zwqxj\") pod \"frr-k8s-ssxwx\" (UID: \"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e\") " pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.557797 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg7td\" (UniqueName: \"kubernetes.io/projected/a20453a0-6ddc-44e5-bd0a-9615f3607127-kube-api-access-zg7td\") pod \"frr-k8s-webhook-server-64bf5d555-48dkt\" (UID: \"a20453a0-6ddc-44e5-bd0a-9615f3607127\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.603499 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611186 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5px75\" (UniqueName: \"kubernetes.io/projected/063f91cf-fe76-4bc6-a070-ccdf43c43f02-kube-api-access-5px75\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611381 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvw4\" (UniqueName: \"kubernetes.io/projected/72277c69-780e-4dcf-a26b-e49001cbf6d1-kube-api-access-pwvw4\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611489 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/063f91cf-fe76-4bc6-a070-ccdf43c43f02-cert\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611584 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611682 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72277c69-780e-4dcf-a26b-e49001cbf6d1-metallb-excludel2\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611794 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/063f91cf-fe76-4bc6-a070-ccdf43c43f02-metrics-certs\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.611887 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-metrics-certs\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.612533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/72277c69-780e-4dcf-a26b-e49001cbf6d1-metallb-excludel2\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: E1004 11:02:32.612647 4758 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 11:02:32 crc kubenswrapper[4758]: E1004 11:02:32.612799 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist podName:72277c69-780e-4dcf-a26b-e49001cbf6d1 nodeName:}" failed. No retries permitted until 2025-10-04 11:02:33.112774716 +0000 UTC m=+730.405425595 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist") pod "speaker-bz5tp" (UID: "72277c69-780e-4dcf-a26b-e49001cbf6d1") : secret "metallb-memberlist" not found Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.613260 4758 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.615199 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.625941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-metrics-certs\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.626711 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/063f91cf-fe76-4bc6-a070-ccdf43c43f02-metrics-certs\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.632710 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/063f91cf-fe76-4bc6-a070-ccdf43c43f02-cert\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.633005 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5px75\" (UniqueName: \"kubernetes.io/projected/063f91cf-fe76-4bc6-a070-ccdf43c43f02-kube-api-access-5px75\") pod \"controller-68d546b9d8-cnb79\" (UID: \"063f91cf-fe76-4bc6-a070-ccdf43c43f02\") " pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.635131 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvw4\" (UniqueName: \"kubernetes.io/projected/72277c69-780e-4dcf-a26b-e49001cbf6d1-kube-api-access-pwvw4\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.737753 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:32 crc kubenswrapper[4758]: I1004 11:02:32.996640 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cnb79"] Oct 04 11:02:33 crc kubenswrapper[4758]: W1004 11:02:33.006060 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod063f91cf_fe76_4bc6_a070_ccdf43c43f02.slice/crio-bbec9ede62f4e011c5311656f0bd3e29d8e76c76425e48854e2b7e81f70be80f WatchSource:0}: Error finding container bbec9ede62f4e011c5311656f0bd3e29d8e76c76425e48854e2b7e81f70be80f: Status 404 returned error can't find the container with id bbec9ede62f4e011c5311656f0bd3e29d8e76c76425e48854e2b7e81f70be80f Oct 04 11:02:33 crc kubenswrapper[4758]: I1004 11:02:33.028507 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt"] Oct 04 11:02:33 crc kubenswrapper[4758]: W1004 11:02:33.034534 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda20453a0_6ddc_44e5_bd0a_9615f3607127.slice/crio-d30ff274c4291738c77b36741c42b3f2baf8d1ec61413123ac9925836695626a WatchSource:0}: Error finding container d30ff274c4291738c77b36741c42b3f2baf8d1ec61413123ac9925836695626a: Status 404 returned error can't find the container with id d30ff274c4291738c77b36741c42b3f2baf8d1ec61413123ac9925836695626a Oct 04 11:02:33 crc kubenswrapper[4758]: I1004 11:02:33.118718 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:33 crc kubenswrapper[4758]: E1004 11:02:33.118884 4758 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 11:02:33 crc kubenswrapper[4758]: E1004 11:02:33.118957 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist podName:72277c69-780e-4dcf-a26b-e49001cbf6d1 nodeName:}" failed. No retries permitted until 2025-10-04 11:02:34.1189404 +0000 UTC m=+731.411591289 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist") pod "speaker-bz5tp" (UID: "72277c69-780e-4dcf-a26b-e49001cbf6d1") : secret "metallb-memberlist" not found Oct 04 11:02:33 crc kubenswrapper[4758]: I1004 11:02:33.150398 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cnb79" event={"ID":"063f91cf-fe76-4bc6-a070-ccdf43c43f02","Type":"ContainerStarted","Data":"aebdd16f5bc0c529625cf6d22eedf02ce3cec7f3081269764e2c1ccc346fbc06"} Oct 04 11:02:33 crc kubenswrapper[4758]: I1004 11:02:33.150435 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cnb79" event={"ID":"063f91cf-fe76-4bc6-a070-ccdf43c43f02","Type":"ContainerStarted","Data":"bbec9ede62f4e011c5311656f0bd3e29d8e76c76425e48854e2b7e81f70be80f"} Oct 04 11:02:33 crc kubenswrapper[4758]: I1004 11:02:33.151177 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" event={"ID":"a20453a0-6ddc-44e5-bd0a-9615f3607127","Type":"ContainerStarted","Data":"d30ff274c4291738c77b36741c42b3f2baf8d1ec61413123ac9925836695626a"} Oct 04 11:02:33 crc kubenswrapper[4758]: I1004 11:02:33.152061 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"ecf09b3691a52228316c63f93f4d96ec9a8b5ad1290a2c0d0ab604f2f0138dae"} Oct 04 11:02:34 crc kubenswrapper[4758]: I1004 11:02:34.132557 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:34 crc kubenswrapper[4758]: I1004 11:02:34.140877 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/72277c69-780e-4dcf-a26b-e49001cbf6d1-memberlist\") pod \"speaker-bz5tp\" (UID: \"72277c69-780e-4dcf-a26b-e49001cbf6d1\") " pod="metallb-system/speaker-bz5tp" Oct 04 11:02:34 crc kubenswrapper[4758]: I1004 11:02:34.159129 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cnb79" event={"ID":"063f91cf-fe76-4bc6-a070-ccdf43c43f02","Type":"ContainerStarted","Data":"6c07dece5f1bc9d3713858ef8e14705ae611cdea23c9142a457df77cc676e2bd"} Oct 04 11:02:34 crc kubenswrapper[4758]: I1004 11:02:34.159313 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:34 crc kubenswrapper[4758]: I1004 11:02:34.227907 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bz5tp" Oct 04 11:02:35 crc kubenswrapper[4758]: I1004 11:02:35.165672 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bz5tp" event={"ID":"72277c69-780e-4dcf-a26b-e49001cbf6d1","Type":"ContainerStarted","Data":"650950196ffa362949fe50ec57e074613f189cdd3fdff084a8d9b3d8f5e9bfb8"} Oct 04 11:02:35 crc kubenswrapper[4758]: I1004 11:02:35.166007 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bz5tp" event={"ID":"72277c69-780e-4dcf-a26b-e49001cbf6d1","Type":"ContainerStarted","Data":"4499d6344143c418d834d9555a4aeead96bd0334807378e3eaff0cdb29d46521"} Oct 04 11:02:35 crc kubenswrapper[4758]: I1004 11:02:35.166017 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bz5tp" event={"ID":"72277c69-780e-4dcf-a26b-e49001cbf6d1","Type":"ContainerStarted","Data":"c746d89c8a8771c10049672df73e6845a04d3c8000ab11833a60d041075d5e7d"} Oct 04 11:02:35 crc kubenswrapper[4758]: I1004 11:02:35.166151 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bz5tp" Oct 04 11:02:35 crc kubenswrapper[4758]: I1004 11:02:35.188689 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-cnb79" podStartSLOduration=3.18867155 podStartE2EDuration="3.18867155s" podCreationTimestamp="2025-10-04 11:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:02:34.185668743 +0000 UTC m=+731.478319632" watchObservedRunningTime="2025-10-04 11:02:35.18867155 +0000 UTC m=+732.481322439" Oct 04 11:02:35 crc kubenswrapper[4758]: I1004 11:02:35.190723 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bz5tp" podStartSLOduration=3.190716197 podStartE2EDuration="3.190716197s" podCreationTimestamp="2025-10-04 11:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:02:35.188800753 +0000 UTC m=+732.481451642" watchObservedRunningTime="2025-10-04 11:02:35.190716197 +0000 UTC m=+732.483367086" Oct 04 11:02:42 crc kubenswrapper[4758]: I1004 11:02:42.225563 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" event={"ID":"a20453a0-6ddc-44e5-bd0a-9615f3607127","Type":"ContainerStarted","Data":"44cb853bfc3b3b74e940a56da18567e378c3d02da722ba57f7ed6d3de8cfbbbf"} Oct 04 11:02:42 crc kubenswrapper[4758]: I1004 11:02:42.228079 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:42 crc kubenswrapper[4758]: I1004 11:02:42.228674 4758 generic.go:334] "Generic (PLEG): container finished" podID="6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e" containerID="f32439685d8e9d0548346ea5a75ebfa55ef4a7d5c5b66a25f96e4c1af9ac48c7" exitCode=0 Oct 04 11:02:42 crc kubenswrapper[4758]: I1004 11:02:42.228736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerDied","Data":"f32439685d8e9d0548346ea5a75ebfa55ef4a7d5c5b66a25f96e4c1af9ac48c7"} Oct 04 11:02:42 crc kubenswrapper[4758]: I1004 11:02:42.309343 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" podStartSLOduration=2.179297924 podStartE2EDuration="10.309319518s" podCreationTimestamp="2025-10-04 11:02:32 +0000 UTC" firstStartedPulling="2025-10-04 11:02:33.036760947 +0000 UTC m=+730.329411836" lastFinishedPulling="2025-10-04 11:02:41.166782541 +0000 UTC m=+738.459433430" observedRunningTime="2025-10-04 11:02:42.25480917 +0000 UTC m=+739.547460079" watchObservedRunningTime="2025-10-04 11:02:42.309319518 +0000 UTC m=+739.601970397" Oct 04 11:02:43 crc kubenswrapper[4758]: I1004 11:02:43.239699 4758 generic.go:334] "Generic (PLEG): container finished" podID="6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e" containerID="90cb040f078bacd8c3bc8315ee7dc79de7d8b6c84dda0770ebb2af8ff13ae0a6" exitCode=0 Oct 04 11:02:43 crc kubenswrapper[4758]: I1004 11:02:43.239818 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerDied","Data":"90cb040f078bacd8c3bc8315ee7dc79de7d8b6c84dda0770ebb2af8ff13ae0a6"} Oct 04 11:02:44 crc kubenswrapper[4758]: I1004 11:02:44.233820 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bz5tp" Oct 04 11:02:44 crc kubenswrapper[4758]: I1004 11:02:44.253651 4758 generic.go:334] "Generic (PLEG): container finished" podID="6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e" containerID="55c3c4dc22a78ce67b296815712ab01b282733c812a84d356c9c3d4d50873030" exitCode=0 Oct 04 11:02:44 crc kubenswrapper[4758]: I1004 11:02:44.253747 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerDied","Data":"55c3c4dc22a78ce67b296815712ab01b282733c812a84d356c9c3d4d50873030"} Oct 04 11:02:45 crc kubenswrapper[4758]: I1004 11:02:45.271420 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"fde29087f1c84550f66074775ebc02d11406d606f8ee3cc61b166e0e9d0931c6"} Oct 04 11:02:45 crc kubenswrapper[4758]: I1004 11:02:45.271884 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"cfe97bcd1bb6329d7d964fc1be78298637fdb329889d1d544f7f66fff55a2101"} Oct 04 11:02:45 crc kubenswrapper[4758]: I1004 11:02:45.271905 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"94bf335a0a0a6394c17ce24893f79bf8f3a003db318f92d5ba903d3ac453ad88"} Oct 04 11:02:45 crc kubenswrapper[4758]: I1004 11:02:45.271922 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"f69479b08bf64eb028ec4b2ec6748740428952e40d4396aa6af3b3bc37cb3bc5"} Oct 04 11:02:45 crc kubenswrapper[4758]: I1004 11:02:45.271939 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"0e2229d6d489798363686a56297def73b5ac8cf6eb441f66c258e093379e95e1"} Oct 04 11:02:46 crc kubenswrapper[4758]: I1004 11:02:46.283429 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-ssxwx" event={"ID":"6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e","Type":"ContainerStarted","Data":"f8d194a99ff97e95c4315db0fc677e18c101d5ed441e0f7eb46e15161f543b5c"} Oct 04 11:02:46 crc kubenswrapper[4758]: I1004 11:02:46.283624 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:46 crc kubenswrapper[4758]: I1004 11:02:46.329850 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-ssxwx" podStartSLOduration=6.041706559 podStartE2EDuration="14.329811492s" podCreationTimestamp="2025-10-04 11:02:32 +0000 UTC" firstStartedPulling="2025-10-04 11:02:32.86056838 +0000 UTC m=+730.153219269" lastFinishedPulling="2025-10-04 11:02:41.148673323 +0000 UTC m=+738.441324202" observedRunningTime="2025-10-04 11:02:46.318178416 +0000 UTC m=+743.610829345" watchObservedRunningTime="2025-10-04 11:02:46.329811492 +0000 UTC m=+743.622462421" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.106329 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-w2kqc"] Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.107191 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.111746 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-czcbl" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.112097 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.112559 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.139218 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w2kqc"] Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.262007 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl5z2\" (UniqueName: \"kubernetes.io/projected/7e988843-c0a2-4466-ad9b-d7ac66f9119e-kube-api-access-gl5z2\") pod \"openstack-operator-index-w2kqc\" (UID: \"7e988843-c0a2-4466-ad9b-d7ac66f9119e\") " pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.363964 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl5z2\" (UniqueName: \"kubernetes.io/projected/7e988843-c0a2-4466-ad9b-d7ac66f9119e-kube-api-access-gl5z2\") pod \"openstack-operator-index-w2kqc\" (UID: \"7e988843-c0a2-4466-ad9b-d7ac66f9119e\") " pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.381908 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl5z2\" (UniqueName: \"kubernetes.io/projected/7e988843-c0a2-4466-ad9b-d7ac66f9119e-kube-api-access-gl5z2\") pod \"openstack-operator-index-w2kqc\" (UID: \"7e988843-c0a2-4466-ad9b-d7ac66f9119e\") " pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.426005 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.604855 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.646133 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:02:47 crc kubenswrapper[4758]: I1004 11:02:47.949487 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-w2kqc"] Oct 04 11:02:47 crc kubenswrapper[4758]: W1004 11:02:47.974776 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e988843_c0a2_4466_ad9b_d7ac66f9119e.slice/crio-01319b31d2bdf3ed6e1ab5fa85f8f5b4be75613b0a9ce2e647f9ef300d249ef3 WatchSource:0}: Error finding container 01319b31d2bdf3ed6e1ab5fa85f8f5b4be75613b0a9ce2e647f9ef300d249ef3: Status 404 returned error can't find the container with id 01319b31d2bdf3ed6e1ab5fa85f8f5b4be75613b0a9ce2e647f9ef300d249ef3 Oct 04 11:02:48 crc kubenswrapper[4758]: I1004 11:02:48.297643 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w2kqc" event={"ID":"7e988843-c0a2-4466-ad9b-d7ac66f9119e","Type":"ContainerStarted","Data":"01319b31d2bdf3ed6e1ab5fa85f8f5b4be75613b0a9ce2e647f9ef300d249ef3"} Oct 04 11:02:50 crc kubenswrapper[4758]: I1004 11:02:50.470755 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-w2kqc"] Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.086047 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t92cq"] Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.088854 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.093780 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t92cq"] Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.220236 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wtzz\" (UniqueName: \"kubernetes.io/projected/6d7e56fa-428b-4d2f-8378-29806b3e66b4-kube-api-access-9wtzz\") pod \"openstack-operator-index-t92cq\" (UID: \"6d7e56fa-428b-4d2f-8378-29806b3e66b4\") " pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.321715 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wtzz\" (UniqueName: \"kubernetes.io/projected/6d7e56fa-428b-4d2f-8378-29806b3e66b4-kube-api-access-9wtzz\") pod \"openstack-operator-index-t92cq\" (UID: \"6d7e56fa-428b-4d2f-8378-29806b3e66b4\") " pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.334538 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-w2kqc" podUID="7e988843-c0a2-4466-ad9b-d7ac66f9119e" containerName="registry-server" containerID="cri-o://e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f" gracePeriod=2 Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.334615 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w2kqc" event={"ID":"7e988843-c0a2-4466-ad9b-d7ac66f9119e","Type":"ContainerStarted","Data":"e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f"} Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.349691 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wtzz\" (UniqueName: \"kubernetes.io/projected/6d7e56fa-428b-4d2f-8378-29806b3e66b4-kube-api-access-9wtzz\") pod \"openstack-operator-index-t92cq\" (UID: \"6d7e56fa-428b-4d2f-8378-29806b3e66b4\") " pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.355206 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-w2kqc" podStartSLOduration=1.569889364 podStartE2EDuration="4.355182395s" podCreationTimestamp="2025-10-04 11:02:47 +0000 UTC" firstStartedPulling="2025-10-04 11:02:47.976073234 +0000 UTC m=+745.268724123" lastFinishedPulling="2025-10-04 11:02:50.761366245 +0000 UTC m=+748.054017154" observedRunningTime="2025-10-04 11:02:51.353298612 +0000 UTC m=+748.645949551" watchObservedRunningTime="2025-10-04 11:02:51.355182395 +0000 UTC m=+748.647833314" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.415042 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.716485 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.828256 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl5z2\" (UniqueName: \"kubernetes.io/projected/7e988843-c0a2-4466-ad9b-d7ac66f9119e-kube-api-access-gl5z2\") pod \"7e988843-c0a2-4466-ad9b-d7ac66f9119e\" (UID: \"7e988843-c0a2-4466-ad9b-d7ac66f9119e\") " Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.832848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e988843-c0a2-4466-ad9b-d7ac66f9119e-kube-api-access-gl5z2" (OuterVolumeSpecName: "kube-api-access-gl5z2") pod "7e988843-c0a2-4466-ad9b-d7ac66f9119e" (UID: "7e988843-c0a2-4466-ad9b-d7ac66f9119e"). InnerVolumeSpecName "kube-api-access-gl5z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.902488 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t92cq"] Oct 04 11:02:51 crc kubenswrapper[4758]: W1004 11:02:51.912522 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d7e56fa_428b_4d2f_8378_29806b3e66b4.slice/crio-668a2a4a4780a8d05d147f5a2894920c0b4555872299e89ebbe5134fb5358ae6 WatchSource:0}: Error finding container 668a2a4a4780a8d05d147f5a2894920c0b4555872299e89ebbe5134fb5358ae6: Status 404 returned error can't find the container with id 668a2a4a4780a8d05d147f5a2894920c0b4555872299e89ebbe5134fb5358ae6 Oct 04 11:02:51 crc kubenswrapper[4758]: I1004 11:02:51.929554 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl5z2\" (UniqueName: \"kubernetes.io/projected/7e988843-c0a2-4466-ad9b-d7ac66f9119e-kube-api-access-gl5z2\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.342383 4758 generic.go:334] "Generic (PLEG): container finished" podID="7e988843-c0a2-4466-ad9b-d7ac66f9119e" containerID="e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f" exitCode=0 Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.342442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w2kqc" event={"ID":"7e988843-c0a2-4466-ad9b-d7ac66f9119e","Type":"ContainerDied","Data":"e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f"} Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.342502 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-w2kqc" event={"ID":"7e988843-c0a2-4466-ad9b-d7ac66f9119e","Type":"ContainerDied","Data":"01319b31d2bdf3ed6e1ab5fa85f8f5b4be75613b0a9ce2e647f9ef300d249ef3"} Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.342534 4758 scope.go:117] "RemoveContainer" containerID="e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f" Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.343226 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-w2kqc" Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.344444 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t92cq" event={"ID":"6d7e56fa-428b-4d2f-8378-29806b3e66b4","Type":"ContainerStarted","Data":"668a2a4a4780a8d05d147f5a2894920c0b4555872299e89ebbe5134fb5358ae6"} Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.370014 4758 scope.go:117] "RemoveContainer" containerID="e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f" Oct 04 11:02:52 crc kubenswrapper[4758]: E1004 11:02:52.370511 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f\": container with ID starting with e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f not found: ID does not exist" containerID="e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f" Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.370541 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f"} err="failed to get container status \"e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f\": rpc error: code = NotFound desc = could not find container \"e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f\": container with ID starting with e4bdb332d1f6c2badf9f048b911ab0e334f0021920f23bc08bc6867df580806f not found: ID does not exist" Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.390511 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-w2kqc"] Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.395501 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-w2kqc"] Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.621258 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-48dkt" Oct 04 11:02:52 crc kubenswrapper[4758]: I1004 11:02:52.745486 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-cnb79" Oct 04 11:02:53 crc kubenswrapper[4758]: I1004 11:02:53.333937 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e988843-c0a2-4466-ad9b-d7ac66f9119e" path="/var/lib/kubelet/pods/7e988843-c0a2-4466-ad9b-d7ac66f9119e/volumes" Oct 04 11:02:53 crc kubenswrapper[4758]: I1004 11:02:53.353113 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t92cq" event={"ID":"6d7e56fa-428b-4d2f-8378-29806b3e66b4","Type":"ContainerStarted","Data":"27d360059df95376375f972a66fb05d258a73da6d6c251dd3f1dd925cef6dc85"} Oct 04 11:02:53 crc kubenswrapper[4758]: I1004 11:02:53.369172 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t92cq" podStartSLOduration=1.754361563 podStartE2EDuration="2.369155592s" podCreationTimestamp="2025-10-04 11:02:51 +0000 UTC" firstStartedPulling="2025-10-04 11:02:51.916064582 +0000 UTC m=+749.208715471" lastFinishedPulling="2025-10-04 11:02:52.530858601 +0000 UTC m=+749.823509500" observedRunningTime="2025-10-04 11:02:53.367931378 +0000 UTC m=+750.660582257" watchObservedRunningTime="2025-10-04 11:02:53.369155592 +0000 UTC m=+750.661806481" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.033245 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8zvst"] Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.033468 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" podUID="83a9a944-0512-4919-b44f-6109a6aa4aa6" containerName="controller-manager" containerID="cri-o://ebc4970a2f2e2b2deaf52766c3adf17759ea58f511635edbae448d2a7b26a386" gracePeriod=30 Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.125446 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c"] Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.125675 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" podUID="d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" containerName="route-controller-manager" containerID="cri-o://1f52a7d951dc8528a34a49e7af719242b46acfe1f9bbaab97b1cc49646199003" gracePeriod=30 Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.367761 4758 generic.go:334] "Generic (PLEG): container finished" podID="83a9a944-0512-4919-b44f-6109a6aa4aa6" containerID="ebc4970a2f2e2b2deaf52766c3adf17759ea58f511635edbae448d2a7b26a386" exitCode=0 Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.367969 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" event={"ID":"83a9a944-0512-4919-b44f-6109a6aa4aa6","Type":"ContainerDied","Data":"ebc4970a2f2e2b2deaf52766c3adf17759ea58f511635edbae448d2a7b26a386"} Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.370642 4758 generic.go:334] "Generic (PLEG): container finished" podID="d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" containerID="1f52a7d951dc8528a34a49e7af719242b46acfe1f9bbaab97b1cc49646199003" exitCode=0 Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.370664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" event={"ID":"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9","Type":"ContainerDied","Data":"1f52a7d951dc8528a34a49e7af719242b46acfe1f9bbaab97b1cc49646199003"} Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.488973 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.524976 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.565504 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-969r6\" (UniqueName: \"kubernetes.io/projected/83a9a944-0512-4919-b44f-6109a6aa4aa6-kube-api-access-969r6\") pod \"83a9a944-0512-4919-b44f-6109a6aa4aa6\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.565645 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-client-ca\") pod \"83a9a944-0512-4919-b44f-6109a6aa4aa6\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.565688 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-config\") pod \"83a9a944-0512-4919-b44f-6109a6aa4aa6\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.565725 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-proxy-ca-bundles\") pod \"83a9a944-0512-4919-b44f-6109a6aa4aa6\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.565820 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a9a944-0512-4919-b44f-6109a6aa4aa6-serving-cert\") pod \"83a9a944-0512-4919-b44f-6109a6aa4aa6\" (UID: \"83a9a944-0512-4919-b44f-6109a6aa4aa6\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.567328 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-client-ca" (OuterVolumeSpecName: "client-ca") pod "83a9a944-0512-4919-b44f-6109a6aa4aa6" (UID: "83a9a944-0512-4919-b44f-6109a6aa4aa6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.567547 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "83a9a944-0512-4919-b44f-6109a6aa4aa6" (UID: "83a9a944-0512-4919-b44f-6109a6aa4aa6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.567894 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-config" (OuterVolumeSpecName: "config") pod "83a9a944-0512-4919-b44f-6109a6aa4aa6" (UID: "83a9a944-0512-4919-b44f-6109a6aa4aa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.571410 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a9a944-0512-4919-b44f-6109a6aa4aa6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "83a9a944-0512-4919-b44f-6109a6aa4aa6" (UID: "83a9a944-0512-4919-b44f-6109a6aa4aa6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.571705 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a9a944-0512-4919-b44f-6109a6aa4aa6-kube-api-access-969r6" (OuterVolumeSpecName: "kube-api-access-969r6") pod "83a9a944-0512-4919-b44f-6109a6aa4aa6" (UID: "83a9a944-0512-4919-b44f-6109a6aa4aa6"). InnerVolumeSpecName "kube-api-access-969r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.666870 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-client-ca\") pod \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.666914 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8stl\" (UniqueName: \"kubernetes.io/projected/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-kube-api-access-b8stl\") pod \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.666950 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-serving-cert\") pod \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.666988 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-config\") pod \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\" (UID: \"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9\") " Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.667185 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a9a944-0512-4919-b44f-6109a6aa4aa6-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.667197 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-969r6\" (UniqueName: \"kubernetes.io/projected/83a9a944-0512-4919-b44f-6109a6aa4aa6-kube-api-access-969r6\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.667206 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.667215 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.667222 4758 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83a9a944-0512-4919-b44f-6109a6aa4aa6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.667832 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-config" (OuterVolumeSpecName: "config") pod "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" (UID: "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.668273 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-client-ca" (OuterVolumeSpecName: "client-ca") pod "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" (UID: "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.671322 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" (UID: "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.671868 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-kube-api-access-b8stl" (OuterVolumeSpecName: "kube-api-access-b8stl") pod "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" (UID: "d3a2e3f5-01b9-43d8-81a2-49c95681b1c9"). InnerVolumeSpecName "kube-api-access-b8stl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.768422 4758 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.768456 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8stl\" (UniqueName: \"kubernetes.io/projected/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-kube-api-access-b8stl\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.768466 4758 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:54 crc kubenswrapper[4758]: I1004 11:02:54.768477 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.378634 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.378731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c" event={"ID":"d3a2e3f5-01b9-43d8-81a2-49c95681b1c9","Type":"ContainerDied","Data":"ce042da7d217ddadaa3c048758d056f6d643483a8c5912f365a593d66b5d25f6"} Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.379074 4758 scope.go:117] "RemoveContainer" containerID="1f52a7d951dc8528a34a49e7af719242b46acfe1f9bbaab97b1cc49646199003" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.380581 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" event={"ID":"83a9a944-0512-4919-b44f-6109a6aa4aa6","Type":"ContainerDied","Data":"76316df3115629407fa3af892b96f6de53781f5636cc9a81aec9d615daf7c3d3"} Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.380637 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8zvst" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.399490 4758 scope.go:117] "RemoveContainer" containerID="ebc4970a2f2e2b2deaf52766c3adf17759ea58f511635edbae448d2a7b26a386" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.407234 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c"] Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.417321 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-l9w7c"] Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.422657 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8zvst"] Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.424966 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8zvst"] Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.445888 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l"] Oct 04 11:02:55 crc kubenswrapper[4758]: E1004 11:02:55.446172 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a9a944-0512-4919-b44f-6109a6aa4aa6" containerName="controller-manager" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446186 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a9a944-0512-4919-b44f-6109a6aa4aa6" containerName="controller-manager" Oct 04 11:02:55 crc kubenswrapper[4758]: E1004 11:02:55.446206 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" containerName="route-controller-manager" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446215 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" containerName="route-controller-manager" Oct 04 11:02:55 crc kubenswrapper[4758]: E1004 11:02:55.446225 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e988843-c0a2-4466-ad9b-d7ac66f9119e" containerName="registry-server" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446233 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e988843-c0a2-4466-ad9b-d7ac66f9119e" containerName="registry-server" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446363 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" containerName="route-controller-manager" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446378 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a9a944-0512-4919-b44f-6109a6aa4aa6" containerName="controller-manager" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446392 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e988843-c0a2-4466-ad9b-d7ac66f9119e" containerName="registry-server" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.446818 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.451378 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.451950 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.453125 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.453469 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.453812 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.456391 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.461444 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l"] Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.577758 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-serving-cert\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.577795 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq2z9\" (UniqueName: \"kubernetes.io/projected/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-kube-api-access-kq2z9\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.577847 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-config\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.577879 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-client-ca\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.678815 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-serving-cert\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.678888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq2z9\" (UniqueName: \"kubernetes.io/projected/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-kube-api-access-kq2z9\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.679555 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-config\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.681325 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-client-ca\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.681540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-config\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.682157 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-client-ca\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.684139 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-serving-cert\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.697727 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq2z9\" (UniqueName: \"kubernetes.io/projected/162cb5be-9f73-43ec-a842-9dd4feeb5fbc-kube-api-access-kq2z9\") pod \"route-controller-manager-8467c6d48d-5lb7l\" (UID: \"162cb5be-9f73-43ec-a842-9dd4feeb5fbc\") " pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:55 crc kubenswrapper[4758]: I1004 11:02:55.760969 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:55.983033 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f894559db-rmwfc"] Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:55.993829 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.002129 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.002305 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.002430 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.002571 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.002706 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.002818 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.003554 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f894559db-rmwfc"] Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.014246 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.061313 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-proxy-ca-bundles\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.061385 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d9be012-85f9-4908-a3fb-b62971a20a32-serving-cert\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.061412 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q47z7\" (UniqueName: \"kubernetes.io/projected/6d9be012-85f9-4908-a3fb-b62971a20a32-kube-api-access-q47z7\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.061448 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-config\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.061467 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-client-ca\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.062223 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l"] Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.163802 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q47z7\" (UniqueName: \"kubernetes.io/projected/6d9be012-85f9-4908-a3fb-b62971a20a32-kube-api-access-q47z7\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.163855 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-config\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.163875 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-client-ca\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.163903 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-proxy-ca-bundles\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.163945 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d9be012-85f9-4908-a3fb-b62971a20a32-serving-cert\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.165467 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-client-ca\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.165761 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-config\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.166664 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d9be012-85f9-4908-a3fb-b62971a20a32-proxy-ca-bundles\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.191030 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q47z7\" (UniqueName: \"kubernetes.io/projected/6d9be012-85f9-4908-a3fb-b62971a20a32-kube-api-access-q47z7\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.194569 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d9be012-85f9-4908-a3fb-b62971a20a32-serving-cert\") pod \"controller-manager-5f894559db-rmwfc\" (UID: \"6d9be012-85f9-4908-a3fb-b62971a20a32\") " pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.367515 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.386669 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" event={"ID":"162cb5be-9f73-43ec-a842-9dd4feeb5fbc","Type":"ContainerStarted","Data":"55c23861e30b10bc9feffbf6120f3a9c39520a6f4bb785b0289b8f30afb5c940"} Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.386705 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" event={"ID":"162cb5be-9f73-43ec-a842-9dd4feeb5fbc","Type":"ContainerStarted","Data":"cf20b6d58cd1b098f2607cfb6db764e11ebddbd6a9942cd7a13271777b527c2e"} Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.387642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.389161 4758 patch_prober.go:28] interesting pod/route-controller-manager-8467c6d48d-5lb7l container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.389233 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" podUID="162cb5be-9f73-43ec-a842-9dd4feeb5fbc" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.413148 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" podStartSLOduration=1.413132402 podStartE2EDuration="1.413132402s" podCreationTimestamp="2025-10-04 11:02:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:02:56.406398593 +0000 UTC m=+753.699049482" watchObservedRunningTime="2025-10-04 11:02:56.413132402 +0000 UTC m=+753.705783291" Oct 04 11:02:56 crc kubenswrapper[4758]: I1004 11:02:56.782826 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f894559db-rmwfc"] Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.332685 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a9a944-0512-4919-b44f-6109a6aa4aa6" path="/var/lib/kubelet/pods/83a9a944-0512-4919-b44f-6109a6aa4aa6/volumes" Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.333400 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a2e3f5-01b9-43d8-81a2-49c95681b1c9" path="/var/lib/kubelet/pods/d3a2e3f5-01b9-43d8-81a2-49c95681b1c9/volumes" Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.414917 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" event={"ID":"6d9be012-85f9-4908-a3fb-b62971a20a32","Type":"ContainerStarted","Data":"536f32019bdf24efb7e8e891344ed5ce0d39129b10b386943d9c8ceca5b345b7"} Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.414958 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" event={"ID":"6d9be012-85f9-4908-a3fb-b62971a20a32","Type":"ContainerStarted","Data":"3190611d79417fd13d1c530047c826a9d113a4dbe70650f1bf9b8f1eebba62a7"} Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.415969 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.420597 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8467c6d48d-5lb7l" Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.424379 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" Oct 04 11:02:57 crc kubenswrapper[4758]: I1004 11:02:57.470269 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f894559db-rmwfc" podStartSLOduration=3.470252634 podStartE2EDuration="3.470252634s" podCreationTimestamp="2025-10-04 11:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:02:57.438018781 +0000 UTC m=+754.730669670" watchObservedRunningTime="2025-10-04 11:02:57.470252634 +0000 UTC m=+754.762903523" Oct 04 11:03:00 crc kubenswrapper[4758]: I1004 11:03:00.530979 4758 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 11:03:01 crc kubenswrapper[4758]: I1004 11:03:01.249430 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:03:01 crc kubenswrapper[4758]: I1004 11:03:01.249777 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:03:01 crc kubenswrapper[4758]: I1004 11:03:01.415372 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:03:01 crc kubenswrapper[4758]: I1004 11:03:01.415420 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:03:01 crc kubenswrapper[4758]: I1004 11:03:01.447056 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:03:01 crc kubenswrapper[4758]: I1004 11:03:01.487089 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-t92cq" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.606389 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-ssxwx" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.711088 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m"] Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.712074 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.715583 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6dwfb" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.735963 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m"] Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.860147 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-util\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.860209 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-bundle\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.860261 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w28xj\" (UniqueName: \"kubernetes.io/projected/a38fe488-755c-40a2-bec7-9606619aff05-kube-api-access-w28xj\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.961758 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w28xj\" (UniqueName: \"kubernetes.io/projected/a38fe488-755c-40a2-bec7-9606619aff05-kube-api-access-w28xj\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.961817 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-util\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.961854 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-bundle\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.962262 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-bundle\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.962476 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-util\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:02 crc kubenswrapper[4758]: I1004 11:03:02.992184 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w28xj\" (UniqueName: \"kubernetes.io/projected/a38fe488-755c-40a2-bec7-9606619aff05-kube-api-access-w28xj\") pod \"bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:03 crc kubenswrapper[4758]: I1004 11:03:03.040785 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:03 crc kubenswrapper[4758]: I1004 11:03:03.512498 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m"] Oct 04 11:03:04 crc kubenswrapper[4758]: I1004 11:03:04.488666 4758 generic.go:334] "Generic (PLEG): container finished" podID="a38fe488-755c-40a2-bec7-9606619aff05" containerID="0ab7ef2dee26880c089d41545d453fc0981c13946a2a911e63b95d9df63f1625" exitCode=0 Oct 04 11:03:04 crc kubenswrapper[4758]: I1004 11:03:04.488732 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" event={"ID":"a38fe488-755c-40a2-bec7-9606619aff05","Type":"ContainerDied","Data":"0ab7ef2dee26880c089d41545d453fc0981c13946a2a911e63b95d9df63f1625"} Oct 04 11:03:04 crc kubenswrapper[4758]: I1004 11:03:04.489085 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" event={"ID":"a38fe488-755c-40a2-bec7-9606619aff05","Type":"ContainerStarted","Data":"8979239cf3c0af0ceff6e13b86daa2e1659a7af346cec8c9c50e59de128ae6d4"} Oct 04 11:03:05 crc kubenswrapper[4758]: I1004 11:03:05.496717 4758 generic.go:334] "Generic (PLEG): container finished" podID="a38fe488-755c-40a2-bec7-9606619aff05" containerID="8b655f176444edf640b0a5fe9e0981979ecf62f6c2b9429ca57425894ae9b40e" exitCode=0 Oct 04 11:03:05 crc kubenswrapper[4758]: I1004 11:03:05.496757 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" event={"ID":"a38fe488-755c-40a2-bec7-9606619aff05","Type":"ContainerDied","Data":"8b655f176444edf640b0a5fe9e0981979ecf62f6c2b9429ca57425894ae9b40e"} Oct 04 11:03:06 crc kubenswrapper[4758]: I1004 11:03:06.504272 4758 generic.go:334] "Generic (PLEG): container finished" podID="a38fe488-755c-40a2-bec7-9606619aff05" containerID="be178d978cd3dc327d7234ac3a1a612f6cf6d75f51b5fd35f14596d3b168d349" exitCode=0 Oct 04 11:03:06 crc kubenswrapper[4758]: I1004 11:03:06.504515 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" event={"ID":"a38fe488-755c-40a2-bec7-9606619aff05","Type":"ContainerDied","Data":"be178d978cd3dc327d7234ac3a1a612f6cf6d75f51b5fd35f14596d3b168d349"} Oct 04 11:03:07 crc kubenswrapper[4758]: I1004 11:03:07.913465 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.025510 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-bundle\") pod \"a38fe488-755c-40a2-bec7-9606619aff05\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.025557 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-util\") pod \"a38fe488-755c-40a2-bec7-9606619aff05\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.025594 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w28xj\" (UniqueName: \"kubernetes.io/projected/a38fe488-755c-40a2-bec7-9606619aff05-kube-api-access-w28xj\") pod \"a38fe488-755c-40a2-bec7-9606619aff05\" (UID: \"a38fe488-755c-40a2-bec7-9606619aff05\") " Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.026521 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-bundle" (OuterVolumeSpecName: "bundle") pod "a38fe488-755c-40a2-bec7-9606619aff05" (UID: "a38fe488-755c-40a2-bec7-9606619aff05"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.038305 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38fe488-755c-40a2-bec7-9606619aff05-kube-api-access-w28xj" (OuterVolumeSpecName: "kube-api-access-w28xj") pod "a38fe488-755c-40a2-bec7-9606619aff05" (UID: "a38fe488-755c-40a2-bec7-9606619aff05"). InnerVolumeSpecName "kube-api-access-w28xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.045093 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-util" (OuterVolumeSpecName: "util") pod "a38fe488-755c-40a2-bec7-9606619aff05" (UID: "a38fe488-755c-40a2-bec7-9606619aff05"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.127871 4758 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.127918 4758 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a38fe488-755c-40a2-bec7-9606619aff05-util\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.127936 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w28xj\" (UniqueName: \"kubernetes.io/projected/a38fe488-755c-40a2-bec7-9606619aff05-kube-api-access-w28xj\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.518637 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" event={"ID":"a38fe488-755c-40a2-bec7-9606619aff05","Type":"ContainerDied","Data":"8979239cf3c0af0ceff6e13b86daa2e1659a7af346cec8c9c50e59de128ae6d4"} Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.518993 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8979239cf3c0af0ceff6e13b86daa2e1659a7af346cec8c9c50e59de128ae6d4" Oct 04 11:03:08 crc kubenswrapper[4758]: I1004 11:03:08.518864 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.087386 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wwbdz"] Oct 04 11:03:09 crc kubenswrapper[4758]: E1004 11:03:09.087765 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="extract" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.087786 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="extract" Oct 04 11:03:09 crc kubenswrapper[4758]: E1004 11:03:09.087821 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="util" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.087835 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="util" Oct 04 11:03:09 crc kubenswrapper[4758]: E1004 11:03:09.087853 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="pull" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.087868 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="pull" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.088056 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38fe488-755c-40a2-bec7-9606619aff05" containerName="extract" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.089502 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.108371 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwbdz"] Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.242673 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-utilities\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.242793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-824rt\" (UniqueName: \"kubernetes.io/projected/73565c31-fdfe-4c8a-894a-3a1b948278ad-kube-api-access-824rt\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.243157 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-catalog-content\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.344221 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-utilities\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.344548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-824rt\" (UniqueName: \"kubernetes.io/projected/73565c31-fdfe-4c8a-894a-3a1b948278ad-kube-api-access-824rt\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.344722 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-catalog-content\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.345078 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-utilities\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.345168 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-catalog-content\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.374138 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-824rt\" (UniqueName: \"kubernetes.io/projected/73565c31-fdfe-4c8a-894a-3a1b948278ad-kube-api-access-824rt\") pod \"redhat-marketplace-wwbdz\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.404536 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:09 crc kubenswrapper[4758]: I1004 11:03:09.809373 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwbdz"] Oct 04 11:03:09 crc kubenswrapper[4758]: W1004 11:03:09.816899 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73565c31_fdfe_4c8a_894a_3a1b948278ad.slice/crio-82ca57dbdf30679dc6117e2cbf906f4f600e4040177de7802eb68fbe2c24450d WatchSource:0}: Error finding container 82ca57dbdf30679dc6117e2cbf906f4f600e4040177de7802eb68fbe2c24450d: Status 404 returned error can't find the container with id 82ca57dbdf30679dc6117e2cbf906f4f600e4040177de7802eb68fbe2c24450d Oct 04 11:03:10 crc kubenswrapper[4758]: I1004 11:03:10.537411 4758 generic.go:334] "Generic (PLEG): container finished" podID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerID="2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218" exitCode=0 Oct 04 11:03:10 crc kubenswrapper[4758]: I1004 11:03:10.537492 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwbdz" event={"ID":"73565c31-fdfe-4c8a-894a-3a1b948278ad","Type":"ContainerDied","Data":"2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218"} Oct 04 11:03:10 crc kubenswrapper[4758]: I1004 11:03:10.539988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwbdz" event={"ID":"73565c31-fdfe-4c8a-894a-3a1b948278ad","Type":"ContainerStarted","Data":"82ca57dbdf30679dc6117e2cbf906f4f600e4040177de7802eb68fbe2c24450d"} Oct 04 11:03:12 crc kubenswrapper[4758]: I1004 11:03:12.554186 4758 generic.go:334] "Generic (PLEG): container finished" podID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerID="094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad" exitCode=0 Oct 04 11:03:12 crc kubenswrapper[4758]: I1004 11:03:12.554274 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwbdz" event={"ID":"73565c31-fdfe-4c8a-894a-3a1b948278ad","Type":"ContainerDied","Data":"094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad"} Oct 04 11:03:13 crc kubenswrapper[4758]: I1004 11:03:13.561972 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwbdz" event={"ID":"73565c31-fdfe-4c8a-894a-3a1b948278ad","Type":"ContainerStarted","Data":"f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57"} Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.066702 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wwbdz" podStartSLOduration=3.53188068 podStartE2EDuration="6.066685871s" podCreationTimestamp="2025-10-04 11:03:09 +0000 UTC" firstStartedPulling="2025-10-04 11:03:10.538920073 +0000 UTC m=+767.831570972" lastFinishedPulling="2025-10-04 11:03:13.073725234 +0000 UTC m=+770.366376163" observedRunningTime="2025-10-04 11:03:13.578477609 +0000 UTC m=+770.871128498" watchObservedRunningTime="2025-10-04 11:03:15.066685871 +0000 UTC m=+772.359336760" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.068976 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r"] Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.069903 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.072057 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-m5zwc" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.088517 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r"] Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.244425 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vmsq\" (UniqueName: \"kubernetes.io/projected/92f9ca3d-6f51-4fcb-917d-dea0d8bea060-kube-api-access-7vmsq\") pod \"openstack-operator-controller-operator-84d79b6cb5-v9k9r\" (UID: \"92f9ca3d-6f51-4fcb-917d-dea0d8bea060\") " pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.346481 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vmsq\" (UniqueName: \"kubernetes.io/projected/92f9ca3d-6f51-4fcb-917d-dea0d8bea060-kube-api-access-7vmsq\") pod \"openstack-operator-controller-operator-84d79b6cb5-v9k9r\" (UID: \"92f9ca3d-6f51-4fcb-917d-dea0d8bea060\") " pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.375866 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vmsq\" (UniqueName: \"kubernetes.io/projected/92f9ca3d-6f51-4fcb-917d-dea0d8bea060-kube-api-access-7vmsq\") pod \"openstack-operator-controller-operator-84d79b6cb5-v9k9r\" (UID: \"92f9ca3d-6f51-4fcb-917d-dea0d8bea060\") " pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.388108 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:15 crc kubenswrapper[4758]: I1004 11:03:15.790965 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r"] Oct 04 11:03:15 crc kubenswrapper[4758]: W1004 11:03:15.806556 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92f9ca3d_6f51_4fcb_917d_dea0d8bea060.slice/crio-76232b12d64e0b6b928b37ddd1624f5fbb359285ef7b1f7bfc594992e1941a7a WatchSource:0}: Error finding container 76232b12d64e0b6b928b37ddd1624f5fbb359285ef7b1f7bfc594992e1941a7a: Status 404 returned error can't find the container with id 76232b12d64e0b6b928b37ddd1624f5fbb359285ef7b1f7bfc594992e1941a7a Oct 04 11:03:16 crc kubenswrapper[4758]: I1004 11:03:16.581552 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" event={"ID":"92f9ca3d-6f51-4fcb-917d-dea0d8bea060","Type":"ContainerStarted","Data":"76232b12d64e0b6b928b37ddd1624f5fbb359285ef7b1f7bfc594992e1941a7a"} Oct 04 11:03:19 crc kubenswrapper[4758]: I1004 11:03:19.405692 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:19 crc kubenswrapper[4758]: I1004 11:03:19.406049 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:19 crc kubenswrapper[4758]: I1004 11:03:19.448854 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:19 crc kubenswrapper[4758]: I1004 11:03:19.638900 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:20 crc kubenswrapper[4758]: I1004 11:03:20.608276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" event={"ID":"92f9ca3d-6f51-4fcb-917d-dea0d8bea060","Type":"ContainerStarted","Data":"fc02c7020b6e901508f58a915653d465cc5d7a2d00357e39111ec3ef52109cc7"} Oct 04 11:03:21 crc kubenswrapper[4758]: I1004 11:03:21.873368 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwbdz"] Oct 04 11:03:21 crc kubenswrapper[4758]: I1004 11:03:21.873689 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wwbdz" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="registry-server" containerID="cri-o://f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57" gracePeriod=2 Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.392311 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.567391 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-utilities\") pod \"73565c31-fdfe-4c8a-894a-3a1b948278ad\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.567475 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-catalog-content\") pod \"73565c31-fdfe-4c8a-894a-3a1b948278ad\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.567567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-824rt\" (UniqueName: \"kubernetes.io/projected/73565c31-fdfe-4c8a-894a-3a1b948278ad-kube-api-access-824rt\") pod \"73565c31-fdfe-4c8a-894a-3a1b948278ad\" (UID: \"73565c31-fdfe-4c8a-894a-3a1b948278ad\") " Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.568356 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-utilities" (OuterVolumeSpecName: "utilities") pod "73565c31-fdfe-4c8a-894a-3a1b948278ad" (UID: "73565c31-fdfe-4c8a-894a-3a1b948278ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.573005 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73565c31-fdfe-4c8a-894a-3a1b948278ad-kube-api-access-824rt" (OuterVolumeSpecName: "kube-api-access-824rt") pod "73565c31-fdfe-4c8a-894a-3a1b948278ad" (UID: "73565c31-fdfe-4c8a-894a-3a1b948278ad"). InnerVolumeSpecName "kube-api-access-824rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.579545 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73565c31-fdfe-4c8a-894a-3a1b948278ad" (UID: "73565c31-fdfe-4c8a-894a-3a1b948278ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.623203 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" event={"ID":"92f9ca3d-6f51-4fcb-917d-dea0d8bea060","Type":"ContainerStarted","Data":"cb07967f2bf6dd301c3fcbe270e1770500602c23e499f9ae3af23ef290a5b730"} Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.623346 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.626089 4758 generic.go:334] "Generic (PLEG): container finished" podID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerID="f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57" exitCode=0 Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.626152 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwbdz" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.626168 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwbdz" event={"ID":"73565c31-fdfe-4c8a-894a-3a1b948278ad","Type":"ContainerDied","Data":"f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57"} Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.626292 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwbdz" event={"ID":"73565c31-fdfe-4c8a-894a-3a1b948278ad","Type":"ContainerDied","Data":"82ca57dbdf30679dc6117e2cbf906f4f600e4040177de7802eb68fbe2c24450d"} Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.626318 4758 scope.go:117] "RemoveContainer" containerID="f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.654699 4758 scope.go:117] "RemoveContainer" containerID="094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.671363 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-824rt\" (UniqueName: \"kubernetes.io/projected/73565c31-fdfe-4c8a-894a-3a1b948278ad-kube-api-access-824rt\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.671395 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.671413 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73565c31-fdfe-4c8a-894a-3a1b948278ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.672943 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" podStartSLOduration=1.400422765 podStartE2EDuration="7.672917927s" podCreationTimestamp="2025-10-04 11:03:15 +0000 UTC" firstStartedPulling="2025-10-04 11:03:15.811676748 +0000 UTC m=+773.104327637" lastFinishedPulling="2025-10-04 11:03:22.08417191 +0000 UTC m=+779.376822799" observedRunningTime="2025-10-04 11:03:22.665340295 +0000 UTC m=+779.957991224" watchObservedRunningTime="2025-10-04 11:03:22.672917927 +0000 UTC m=+779.965568816" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.690740 4758 scope.go:117] "RemoveContainer" containerID="2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.692378 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwbdz"] Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.696500 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwbdz"] Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.710147 4758 scope.go:117] "RemoveContainer" containerID="f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57" Oct 04 11:03:22 crc kubenswrapper[4758]: E1004 11:03:22.710563 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57\": container with ID starting with f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57 not found: ID does not exist" containerID="f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.710611 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57"} err="failed to get container status \"f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57\": rpc error: code = NotFound desc = could not find container \"f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57\": container with ID starting with f6cbfc0b3ee96719a608081a79f132800b00e034c7c8b2719a9a69d8292eee57 not found: ID does not exist" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.710638 4758 scope.go:117] "RemoveContainer" containerID="094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad" Oct 04 11:03:22 crc kubenswrapper[4758]: E1004 11:03:22.711056 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad\": container with ID starting with 094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad not found: ID does not exist" containerID="094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.711097 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad"} err="failed to get container status \"094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad\": rpc error: code = NotFound desc = could not find container \"094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad\": container with ID starting with 094f6ffa53ce63fea37802677ec60f47ca97efc7dd890eb6a27da7d3a3f744ad not found: ID does not exist" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.711135 4758 scope.go:117] "RemoveContainer" containerID="2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218" Oct 04 11:03:22 crc kubenswrapper[4758]: E1004 11:03:22.711461 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218\": container with ID starting with 2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218 not found: ID does not exist" containerID="2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218" Oct 04 11:03:22 crc kubenswrapper[4758]: I1004 11:03:22.711489 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218"} err="failed to get container status \"2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218\": rpc error: code = NotFound desc = could not find container \"2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218\": container with ID starting with 2e7d85808a7328b9998ba45e2adcab6b3e4639f07bd5862ae0802800b1b3e218 not found: ID does not exist" Oct 04 11:03:23 crc kubenswrapper[4758]: I1004 11:03:23.338369 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" path="/var/lib/kubelet/pods/73565c31-fdfe-4c8a-894a-3a1b948278ad/volumes" Oct 04 11:03:25 crc kubenswrapper[4758]: I1004 11:03:25.391881 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-84d79b6cb5-v9k9r" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.475699 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-55trd"] Oct 04 11:03:26 crc kubenswrapper[4758]: E1004 11:03:26.475908 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="extract-utilities" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.475920 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="extract-utilities" Oct 04 11:03:26 crc kubenswrapper[4758]: E1004 11:03:26.475930 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="extract-content" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.475936 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="extract-content" Oct 04 11:03:26 crc kubenswrapper[4758]: E1004 11:03:26.475951 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="registry-server" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.475957 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="registry-server" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.476053 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="73565c31-fdfe-4c8a-894a-3a1b948278ad" containerName="registry-server" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.476804 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.522634 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-catalog-content\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.522681 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrghr\" (UniqueName: \"kubernetes.io/projected/a85a5740-14a8-41c7-b319-b44ae267a0e2-kube-api-access-vrghr\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.522696 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-utilities\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.524548 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-55trd"] Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.623678 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-catalog-content\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.623732 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrghr\" (UniqueName: \"kubernetes.io/projected/a85a5740-14a8-41c7-b319-b44ae267a0e2-kube-api-access-vrghr\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.623749 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-utilities\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.624203 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-utilities\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.624267 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-catalog-content\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.653801 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrghr\" (UniqueName: \"kubernetes.io/projected/a85a5740-14a8-41c7-b319-b44ae267a0e2-kube-api-access-vrghr\") pod \"certified-operators-55trd\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:26 crc kubenswrapper[4758]: I1004 11:03:26.793306 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:27 crc kubenswrapper[4758]: I1004 11:03:27.259524 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-55trd"] Oct 04 11:03:27 crc kubenswrapper[4758]: I1004 11:03:27.663667 4758 generic.go:334] "Generic (PLEG): container finished" podID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerID="06616d902cece10978016d3d91c012c551951f5d854947de2f55e00237dff7b9" exitCode=0 Oct 04 11:03:27 crc kubenswrapper[4758]: I1004 11:03:27.663702 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerDied","Data":"06616d902cece10978016d3d91c012c551951f5d854947de2f55e00237dff7b9"} Oct 04 11:03:27 crc kubenswrapper[4758]: I1004 11:03:27.663725 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerStarted","Data":"b965cf0c23f21e35de8c2e49cf4c236442197d78dbba2d99ed3266f34aa0e555"} Oct 04 11:03:28 crc kubenswrapper[4758]: I1004 11:03:28.678992 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerStarted","Data":"89a8a28f9637bf3af25811a587c03c6836f50b3f04542fd884ea0c6518e291e6"} Oct 04 11:03:29 crc kubenswrapper[4758]: I1004 11:03:29.688646 4758 generic.go:334] "Generic (PLEG): container finished" podID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerID="89a8a28f9637bf3af25811a587c03c6836f50b3f04542fd884ea0c6518e291e6" exitCode=0 Oct 04 11:03:29 crc kubenswrapper[4758]: I1004 11:03:29.688710 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerDied","Data":"89a8a28f9637bf3af25811a587c03c6836f50b3f04542fd884ea0c6518e291e6"} Oct 04 11:03:30 crc kubenswrapper[4758]: I1004 11:03:30.696761 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerStarted","Data":"821af43a30801ef46d0fec1cc7fbfc3fc93315aece43d491102ac033d8365e51"} Oct 04 11:03:30 crc kubenswrapper[4758]: I1004 11:03:30.714920 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-55trd" podStartSLOduration=2.284815097 podStartE2EDuration="4.714906144s" podCreationTimestamp="2025-10-04 11:03:26 +0000 UTC" firstStartedPulling="2025-10-04 11:03:27.665611135 +0000 UTC m=+784.958262044" lastFinishedPulling="2025-10-04 11:03:30.095702202 +0000 UTC m=+787.388353091" observedRunningTime="2025-10-04 11:03:30.714632546 +0000 UTC m=+788.007283445" watchObservedRunningTime="2025-10-04 11:03:30.714906144 +0000 UTC m=+788.007557033" Oct 04 11:03:31 crc kubenswrapper[4758]: I1004 11:03:31.249324 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:03:31 crc kubenswrapper[4758]: I1004 11:03:31.249572 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:03:36 crc kubenswrapper[4758]: I1004 11:03:36.794172 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:36 crc kubenswrapper[4758]: I1004 11:03:36.794798 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:36 crc kubenswrapper[4758]: I1004 11:03:36.871186 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:37 crc kubenswrapper[4758]: I1004 11:03:37.809852 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:37 crc kubenswrapper[4758]: I1004 11:03:37.867467 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-55trd"] Oct 04 11:03:39 crc kubenswrapper[4758]: I1004 11:03:39.764697 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-55trd" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="registry-server" containerID="cri-o://821af43a30801ef46d0fec1cc7fbfc3fc93315aece43d491102ac033d8365e51" gracePeriod=2 Oct 04 11:03:40 crc kubenswrapper[4758]: I1004 11:03:40.771896 4758 generic.go:334] "Generic (PLEG): container finished" podID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerID="821af43a30801ef46d0fec1cc7fbfc3fc93315aece43d491102ac033d8365e51" exitCode=0 Oct 04 11:03:40 crc kubenswrapper[4758]: I1004 11:03:40.771988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerDied","Data":"821af43a30801ef46d0fec1cc7fbfc3fc93315aece43d491102ac033d8365e51"} Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.314326 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.508524 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-utilities\") pod \"a85a5740-14a8-41c7-b319-b44ae267a0e2\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.508622 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrghr\" (UniqueName: \"kubernetes.io/projected/a85a5740-14a8-41c7-b319-b44ae267a0e2-kube-api-access-vrghr\") pod \"a85a5740-14a8-41c7-b319-b44ae267a0e2\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.508663 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-catalog-content\") pod \"a85a5740-14a8-41c7-b319-b44ae267a0e2\" (UID: \"a85a5740-14a8-41c7-b319-b44ae267a0e2\") " Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.509384 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-utilities" (OuterVolumeSpecName: "utilities") pod "a85a5740-14a8-41c7-b319-b44ae267a0e2" (UID: "a85a5740-14a8-41c7-b319-b44ae267a0e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.519850 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85a5740-14a8-41c7-b319-b44ae267a0e2-kube-api-access-vrghr" (OuterVolumeSpecName: "kube-api-access-vrghr") pod "a85a5740-14a8-41c7-b319-b44ae267a0e2" (UID: "a85a5740-14a8-41c7-b319-b44ae267a0e2"). InnerVolumeSpecName "kube-api-access-vrghr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.562070 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a85a5740-14a8-41c7-b319-b44ae267a0e2" (UID: "a85a5740-14a8-41c7-b319-b44ae267a0e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.609711 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrghr\" (UniqueName: \"kubernetes.io/projected/a85a5740-14a8-41c7-b319-b44ae267a0e2-kube-api-access-vrghr\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.609748 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.609758 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a85a5740-14a8-41c7-b319-b44ae267a0e2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.780051 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55trd" event={"ID":"a85a5740-14a8-41c7-b319-b44ae267a0e2","Type":"ContainerDied","Data":"b965cf0c23f21e35de8c2e49cf4c236442197d78dbba2d99ed3266f34aa0e555"} Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.780126 4758 scope.go:117] "RemoveContainer" containerID="821af43a30801ef46d0fec1cc7fbfc3fc93315aece43d491102ac033d8365e51" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.780306 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55trd" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.805282 4758 scope.go:117] "RemoveContainer" containerID="89a8a28f9637bf3af25811a587c03c6836f50b3f04542fd884ea0c6518e291e6" Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.826223 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-55trd"] Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.834136 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-55trd"] Oct 04 11:03:41 crc kubenswrapper[4758]: I1004 11:03:41.838775 4758 scope.go:117] "RemoveContainer" containerID="06616d902cece10978016d3d91c012c551951f5d854947de2f55e00237dff7b9" Oct 04 11:03:43 crc kubenswrapper[4758]: I1004 11:03:43.336062 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" path="/var/lib/kubelet/pods/a85a5740-14a8-41c7-b319-b44ae267a0e2/volumes" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.844550 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn"] Oct 04 11:03:47 crc kubenswrapper[4758]: E1004 11:03:47.845406 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="extract-content" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.845425 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="extract-content" Oct 04 11:03:47 crc kubenswrapper[4758]: E1004 11:03:47.845449 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="extract-utilities" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.845457 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="extract-utilities" Oct 04 11:03:47 crc kubenswrapper[4758]: E1004 11:03:47.845468 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="registry-server" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.845474 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="registry-server" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.845596 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85a5740-14a8-41c7-b319-b44ae267a0e2" containerName="registry-server" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.846205 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.847928 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.848589 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hwzvb" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.848696 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.850586 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jcsld" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.867923 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.871369 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.878722 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.879707 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.885295 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9899b" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.901408 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.907810 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfx45\" (UniqueName: \"kubernetes.io/projected/6c293821-0ef9-4bf3-9a2a-f188ac3b2151-kube-api-access-qfx45\") pod \"barbican-operator-controller-manager-7d49d7896f-mgphn\" (UID: \"6c293821-0ef9-4bf3-9a2a-f188ac3b2151\") " pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.907886 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phh5l\" (UniqueName: \"kubernetes.io/projected/59448e96-d587-493a-b5c5-897aa1167152-kube-api-access-phh5l\") pod \"designate-operator-controller-manager-5d8bb764f9-77mnr\" (UID: \"59448e96-d587-493a-b5c5-897aa1167152\") " pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.907918 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqwgb\" (UniqueName: \"kubernetes.io/projected/112ea09c-f473-4dbc-85d6-1e0ef7e6cf62-kube-api-access-fqwgb\") pod \"cinder-operator-controller-manager-866c4585d9-frqpq\" (UID: \"112ea09c-f473-4dbc-85d6-1e0ef7e6cf62\") " pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.928883 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.929817 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.933012 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mqvf4" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.935197 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.941865 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.943530 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.945829 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vvljd" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.963659 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.972308 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls"] Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.976874 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.988414 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-j25vr" Oct 04 11:03:47 crc kubenswrapper[4758]: I1004 11:03:47.991499 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.004523 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.005500 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.009020 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phh5l\" (UniqueName: \"kubernetes.io/projected/59448e96-d587-493a-b5c5-897aa1167152-kube-api-access-phh5l\") pod \"designate-operator-controller-manager-5d8bb764f9-77mnr\" (UID: \"59448e96-d587-493a-b5c5-897aa1167152\") " pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.009060 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvc28\" (UniqueName: \"kubernetes.io/projected/e261e6dd-0cdb-43b4-8981-5c34a7288fb3-kube-api-access-bvc28\") pod \"glance-operator-controller-manager-6b95d97c7f-lbtmk\" (UID: \"e261e6dd-0cdb-43b4-8981-5c34a7288fb3\") " pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.009094 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqwgb\" (UniqueName: \"kubernetes.io/projected/112ea09c-f473-4dbc-85d6-1e0ef7e6cf62-kube-api-access-fqwgb\") pod \"cinder-operator-controller-manager-866c4585d9-frqpq\" (UID: \"112ea09c-f473-4dbc-85d6-1e0ef7e6cf62\") " pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.009135 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppxs7\" (UniqueName: \"kubernetes.io/projected/f22d4cbc-37ae-4621-bc6d-4818642ecd29-kube-api-access-ppxs7\") pod \"horizon-operator-controller-manager-dd56696bb-dm2ls\" (UID: \"f22d4cbc-37ae-4621-bc6d-4818642ecd29\") " pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.009167 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj9h5\" (UniqueName: \"kubernetes.io/projected/30c739ae-3a42-42c9-babb-03a8dd6b1a50-kube-api-access-lj9h5\") pod \"heat-operator-controller-manager-697cc69689-gmq9x\" (UID: \"30c739ae-3a42-42c9-babb-03a8dd6b1a50\") " pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.009188 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfx45\" (UniqueName: \"kubernetes.io/projected/6c293821-0ef9-4bf3-9a2a-f188ac3b2151-kube-api-access-qfx45\") pod \"barbican-operator-controller-manager-7d49d7896f-mgphn\" (UID: \"6c293821-0ef9-4bf3-9a2a-f188ac3b2151\") " pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.014601 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.015549 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.015966 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ckpvf" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.018976 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.021695 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jpk6h" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.046051 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.074321 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfx45\" (UniqueName: \"kubernetes.io/projected/6c293821-0ef9-4bf3-9a2a-f188ac3b2151-kube-api-access-qfx45\") pod \"barbican-operator-controller-manager-7d49d7896f-mgphn\" (UID: \"6c293821-0ef9-4bf3-9a2a-f188ac3b2151\") " pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.076355 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phh5l\" (UniqueName: \"kubernetes.io/projected/59448e96-d587-493a-b5c5-897aa1167152-kube-api-access-phh5l\") pod \"designate-operator-controller-manager-5d8bb764f9-77mnr\" (UID: \"59448e96-d587-493a-b5c5-897aa1167152\") " pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.081202 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.102595 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqwgb\" (UniqueName: \"kubernetes.io/projected/112ea09c-f473-4dbc-85d6-1e0ef7e6cf62-kube-api-access-fqwgb\") pod \"cinder-operator-controller-manager-866c4585d9-frqpq\" (UID: \"112ea09c-f473-4dbc-85d6-1e0ef7e6cf62\") " pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.108161 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.109200 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.110363 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvc28\" (UniqueName: \"kubernetes.io/projected/e261e6dd-0cdb-43b4-8981-5c34a7288fb3-kube-api-access-bvc28\") pod \"glance-operator-controller-manager-6b95d97c7f-lbtmk\" (UID: \"e261e6dd-0cdb-43b4-8981-5c34a7288fb3\") " pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.110411 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qbnl\" (UniqueName: \"kubernetes.io/projected/2bafb624-fe21-45cb-b787-192d4353378c-kube-api-access-8qbnl\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.110442 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppxs7\" (UniqueName: \"kubernetes.io/projected/f22d4cbc-37ae-4621-bc6d-4818642ecd29-kube-api-access-ppxs7\") pod \"horizon-operator-controller-manager-dd56696bb-dm2ls\" (UID: \"f22d4cbc-37ae-4621-bc6d-4818642ecd29\") " pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.110471 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clr4m\" (UniqueName: \"kubernetes.io/projected/bfb96416-cb87-48e2-9ea7-71a899c3a620-kube-api-access-clr4m\") pod \"ironic-operator-controller-manager-578455559c-l7m4w\" (UID: \"bfb96416-cb87-48e2-9ea7-71a899c3a620\") " pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.110499 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.110517 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj9h5\" (UniqueName: \"kubernetes.io/projected/30c739ae-3a42-42c9-babb-03a8dd6b1a50-kube-api-access-lj9h5\") pod \"heat-operator-controller-manager-697cc69689-gmq9x\" (UID: \"30c739ae-3a42-42c9-babb-03a8dd6b1a50\") " pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.113862 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.116058 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-qsbmx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.162751 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppxs7\" (UniqueName: \"kubernetes.io/projected/f22d4cbc-37ae-4621-bc6d-4818642ecd29-kube-api-access-ppxs7\") pod \"horizon-operator-controller-manager-dd56696bb-dm2ls\" (UID: \"f22d4cbc-37ae-4621-bc6d-4818642ecd29\") " pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.163400 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.164502 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.165664 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj9h5\" (UniqueName: \"kubernetes.io/projected/30c739ae-3a42-42c9-babb-03a8dd6b1a50-kube-api-access-lj9h5\") pod \"heat-operator-controller-manager-697cc69689-gmq9x\" (UID: \"30c739ae-3a42-42c9-babb-03a8dd6b1a50\") " pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.173695 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-zdkth" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.178968 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.190392 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvc28\" (UniqueName: \"kubernetes.io/projected/e261e6dd-0cdb-43b4-8981-5c34a7288fb3-kube-api-access-bvc28\") pod \"glance-operator-controller-manager-6b95d97c7f-lbtmk\" (UID: \"e261e6dd-0cdb-43b4-8981-5c34a7288fb3\") " pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.195342 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.208758 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.211261 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clr4m\" (UniqueName: \"kubernetes.io/projected/bfb96416-cb87-48e2-9ea7-71a899c3a620-kube-api-access-clr4m\") pod \"ironic-operator-controller-manager-578455559c-l7m4w\" (UID: \"bfb96416-cb87-48e2-9ea7-71a899c3a620\") " pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.211293 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.211318 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmvxm\" (UniqueName: \"kubernetes.io/projected/aa921475-1667-4eff-8553-9a38928cf00f-kube-api-access-rmvxm\") pod \"keystone-operator-controller-manager-795ccf785c-jtcj2\" (UID: \"aa921475-1667-4eff-8553-9a38928cf00f\") " pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.211369 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-285n6\" (UniqueName: \"kubernetes.io/projected/95aa14ab-a652-4eb4-97c1-7ab47431913b-kube-api-access-285n6\") pod \"manila-operator-controller-manager-6679c59c97-r7gpj\" (UID: \"95aa14ab-a652-4eb4-97c1-7ab47431913b\") " pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.211395 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qbnl\" (UniqueName: \"kubernetes.io/projected/2bafb624-fe21-45cb-b787-192d4353378c-kube-api-access-8qbnl\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.219657 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.219722 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert podName:2bafb624-fe21-45cb-b787-192d4353378c nodeName:}" failed. No retries permitted until 2025-10-04 11:03:48.719703952 +0000 UTC m=+806.012354831 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert") pod "infra-operator-controller-manager-95c44c5bb-kg2lj" (UID: "2bafb624-fe21-45cb-b787-192d4353378c") : secret "infra-operator-webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.219881 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.245495 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.246510 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.249314 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-prjgx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.259048 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qbnl\" (UniqueName: \"kubernetes.io/projected/2bafb624-fe21-45cb-b787-192d4353378c-kube-api-access-8qbnl\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.263301 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.265880 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clr4m\" (UniqueName: \"kubernetes.io/projected/bfb96416-cb87-48e2-9ea7-71a899c3a620-kube-api-access-clr4m\") pod \"ironic-operator-controller-manager-578455559c-l7m4w\" (UID: \"bfb96416-cb87-48e2-9ea7-71a899c3a620\") " pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.269414 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.289045 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.297228 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.299738 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.302090 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-95v82" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.307970 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.328529 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-285n6\" (UniqueName: \"kubernetes.io/projected/95aa14ab-a652-4eb4-97c1-7ab47431913b-kube-api-access-285n6\") pod \"manila-operator-controller-manager-6679c59c97-r7gpj\" (UID: \"95aa14ab-a652-4eb4-97c1-7ab47431913b\") " pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.328572 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9gts\" (UniqueName: \"kubernetes.io/projected/7950908c-ddc0-434f-b28a-8d8687392a53-kube-api-access-v9gts\") pod \"mariadb-operator-controller-manager-746ff9ccdc-d269t\" (UID: \"7950908c-ddc0-434f-b28a-8d8687392a53\") " pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.328651 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmvxm\" (UniqueName: \"kubernetes.io/projected/aa921475-1667-4eff-8553-9a38928cf00f-kube-api-access-rmvxm\") pod \"keystone-operator-controller-manager-795ccf785c-jtcj2\" (UID: \"aa921475-1667-4eff-8553-9a38928cf00f\") " pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.328678 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ccls\" (UniqueName: \"kubernetes.io/projected/1f16ce8c-ba77-4c4e-8f48-ee981b138a01-kube-api-access-4ccls\") pod \"neutron-operator-controller-manager-74d4d6c8f6-7ndwj\" (UID: \"1f16ce8c-ba77-4c4e-8f48-ee981b138a01\") " pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.348293 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.351271 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.363690 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmvxm\" (UniqueName: \"kubernetes.io/projected/aa921475-1667-4eff-8553-9a38928cf00f-kube-api-access-rmvxm\") pod \"keystone-operator-controller-manager-795ccf785c-jtcj2\" (UID: \"aa921475-1667-4eff-8553-9a38928cf00f\") " pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.364756 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-285n6\" (UniqueName: \"kubernetes.io/projected/95aa14ab-a652-4eb4-97c1-7ab47431913b-kube-api-access-285n6\") pod \"manila-operator-controller-manager-6679c59c97-r7gpj\" (UID: \"95aa14ab-a652-4eb4-97c1-7ab47431913b\") " pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.403298 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.406224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.420341 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-h52rv" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.436209 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.436982 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ccls\" (UniqueName: \"kubernetes.io/projected/1f16ce8c-ba77-4c4e-8f48-ee981b138a01-kube-api-access-4ccls\") pod \"neutron-operator-controller-manager-74d4d6c8f6-7ndwj\" (UID: \"1f16ce8c-ba77-4c4e-8f48-ee981b138a01\") " pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.437016 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzzx2\" (UniqueName: \"kubernetes.io/projected/66ed67ad-7738-4701-9e08-0e45b5ac73d2-kube-api-access-wzzx2\") pod \"nova-operator-controller-manager-57b88b6ccd-m24qr\" (UID: \"66ed67ad-7738-4701-9e08-0e45b5ac73d2\") " pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.437062 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9gts\" (UniqueName: \"kubernetes.io/projected/7950908c-ddc0-434f-b28a-8d8687392a53-kube-api-access-v9gts\") pod \"mariadb-operator-controller-manager-746ff9ccdc-d269t\" (UID: \"7950908c-ddc0-434f-b28a-8d8687392a53\") " pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.466082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9gts\" (UniqueName: \"kubernetes.io/projected/7950908c-ddc0-434f-b28a-8d8687392a53-kube-api-access-v9gts\") pod \"mariadb-operator-controller-manager-746ff9ccdc-d269t\" (UID: \"7950908c-ddc0-434f-b28a-8d8687392a53\") " pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.473378 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ccls\" (UniqueName: \"kubernetes.io/projected/1f16ce8c-ba77-4c4e-8f48-ee981b138a01-kube-api-access-4ccls\") pod \"neutron-operator-controller-manager-74d4d6c8f6-7ndwj\" (UID: \"1f16ce8c-ba77-4c4e-8f48-ee981b138a01\") " pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.488456 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.488503 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.489948 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.493433 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-7dssz" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.498387 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.513183 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.514274 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.518364 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5tcmc" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.518511 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.539696 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzzx2\" (UniqueName: \"kubernetes.io/projected/66ed67ad-7738-4701-9e08-0e45b5ac73d2-kube-api-access-wzzx2\") pod \"nova-operator-controller-manager-57b88b6ccd-m24qr\" (UID: \"66ed67ad-7738-4701-9e08-0e45b5ac73d2\") " pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.539734 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rwzs\" (UniqueName: \"kubernetes.io/projected/4356f9f3-4e48-4ca8-92d4-f55914ded96b-kube-api-access-2rwzs\") pod \"octavia-operator-controller-manager-768ddcd459-7zwcs\" (UID: \"4356f9f3-4e48-4ca8-92d4-f55914ded96b\") " pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.539769 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmq76\" (UniqueName: \"kubernetes.io/projected/5b89709e-0a88-4e46-86e6-09e5b1b604da-kube-api-access-dmq76\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.539787 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b89709e-0a88-4e46-86e6-09e5b1b604da-cert\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.552435 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.559795 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.578307 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.579137 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.583610 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-sbhlj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.621165 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzzx2\" (UniqueName: \"kubernetes.io/projected/66ed67ad-7738-4701-9e08-0e45b5ac73d2-kube-api-access-wzzx2\") pod \"nova-operator-controller-manager-57b88b6ccd-m24qr\" (UID: \"66ed67ad-7738-4701-9e08-0e45b5ac73d2\") " pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.622246 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.623294 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.625184 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-v662q" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.629223 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.634544 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.640926 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b89709e-0a88-4e46-86e6-09e5b1b604da-cert\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.640981 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmq76\" (UniqueName: \"kubernetes.io/projected/5b89709e-0a88-4e46-86e6-09e5b1b604da-kube-api-access-dmq76\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.641053 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54q5z\" (UniqueName: \"kubernetes.io/projected/dd449335-3579-4bea-a408-e8b92538f3fa-kube-api-access-54q5z\") pod \"placement-operator-controller-manager-5b97cccc57-g8m77\" (UID: \"dd449335-3579-4bea-a408-e8b92538f3fa\") " pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.641117 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c92h\" (UniqueName: \"kubernetes.io/projected/cc6e56f4-7374-4ffb-b5f2-55525602b161-kube-api-access-6c92h\") pod \"ovn-operator-controller-manager-d879db64d-r49wx\" (UID: \"cc6e56f4-7374-4ffb-b5f2-55525602b161\") " pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.641164 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rwzs\" (UniqueName: \"kubernetes.io/projected/4356f9f3-4e48-4ca8-92d4-f55914ded96b-kube-api-access-2rwzs\") pod \"octavia-operator-controller-manager-768ddcd459-7zwcs\" (UID: \"4356f9f3-4e48-4ca8-92d4-f55914ded96b\") " pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.641596 4758 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.641641 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b89709e-0a88-4e46-86e6-09e5b1b604da-cert podName:5b89709e-0a88-4e46-86e6-09e5b1b604da nodeName:}" failed. No retries permitted until 2025-10-04 11:03:49.141628835 +0000 UTC m=+806.434279714 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b89709e-0a88-4e46-86e6-09e5b1b604da-cert") pod "openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" (UID: "5b89709e-0a88-4e46-86e6-09e5b1b604da") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.641952 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.649338 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.650764 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.656849 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.659894 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.667796 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wcv87" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.671402 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.671719 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-sn5lh" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.690744 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.691869 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.696928 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-qtnjc" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.704261 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.711231 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rwzs\" (UniqueName: \"kubernetes.io/projected/4356f9f3-4e48-4ca8-92d4-f55914ded96b-kube-api-access-2rwzs\") pod \"octavia-operator-controller-manager-768ddcd459-7zwcs\" (UID: \"4356f9f3-4e48-4ca8-92d4-f55914ded96b\") " pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.713893 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.725144 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmq76\" (UniqueName: \"kubernetes.io/projected/5b89709e-0a88-4e46-86e6-09e5b1b604da-kube-api-access-dmq76\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.742048 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwb58\" (UniqueName: \"kubernetes.io/projected/bb6211ba-a514-4243-9204-0eb8954431cd-kube-api-access-nwb58\") pod \"swift-operator-controller-manager-7db4cd47d4-gmj7n\" (UID: \"bb6211ba-a514-4243-9204-0eb8954431cd\") " pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.742093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54q5z\" (UniqueName: \"kubernetes.io/projected/dd449335-3579-4bea-a408-e8b92538f3fa-kube-api-access-54q5z\") pod \"placement-operator-controller-manager-5b97cccc57-g8m77\" (UID: \"dd449335-3579-4bea-a408-e8b92538f3fa\") " pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.742130 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z85dd\" (UniqueName: \"kubernetes.io/projected/d9e18e05-fb58-4fa9-b4be-be28657d52ee-kube-api-access-z85dd\") pod \"telemetry-operator-controller-manager-5b7d9bd964-gqwrj\" (UID: \"d9e18e05-fb58-4fa9-b4be-be28657d52ee\") " pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.742176 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7kpp\" (UniqueName: \"kubernetes.io/projected/5d2ba7ed-1581-4cc2-957d-42ee5bf33781-kube-api-access-r7kpp\") pod \"test-operator-controller-manager-5f4f4847c9-gpsls\" (UID: \"5d2ba7ed-1581-4cc2-957d-42ee5bf33781\") " pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.742304 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c92h\" (UniqueName: \"kubernetes.io/projected/cc6e56f4-7374-4ffb-b5f2-55525602b161-kube-api-access-6c92h\") pod \"ovn-operator-controller-manager-d879db64d-r49wx\" (UID: \"cc6e56f4-7374-4ffb-b5f2-55525602b161\") " pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.742385 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.742691 4758 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.742741 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert podName:2bafb624-fe21-45cb-b787-192d4353378c nodeName:}" failed. No retries permitted until 2025-10-04 11:03:49.742724838 +0000 UTC m=+807.035375727 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert") pod "infra-operator-controller-manager-95c44c5bb-kg2lj" (UID: "2bafb624-fe21-45cb-b787-192d4353378c") : secret "infra-operator-webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.751166 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.762520 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.775998 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54q5z\" (UniqueName: \"kubernetes.io/projected/dd449335-3579-4bea-a408-e8b92538f3fa-kube-api-access-54q5z\") pod \"placement-operator-controller-manager-5b97cccc57-g8m77\" (UID: \"dd449335-3579-4bea-a408-e8b92538f3fa\") " pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.776914 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c92h\" (UniqueName: \"kubernetes.io/projected/cc6e56f4-7374-4ffb-b5f2-55525602b161-kube-api-access-6c92h\") pod \"ovn-operator-controller-manager-d879db64d-r49wx\" (UID: \"cc6e56f4-7374-4ffb-b5f2-55525602b161\") " pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.808644 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.816536 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.817943 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.820148 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.820294 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ffxm5" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.843892 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzrb\" (UniqueName: \"kubernetes.io/projected/93fd4e69-37c4-485a-b780-6e366ebd0f65-kube-api-access-2qzrb\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.843942 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwb58\" (UniqueName: \"kubernetes.io/projected/bb6211ba-a514-4243-9204-0eb8954431cd-kube-api-access-nwb58\") pod \"swift-operator-controller-manager-7db4cd47d4-gmj7n\" (UID: \"bb6211ba-a514-4243-9204-0eb8954431cd\") " pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.843982 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z85dd\" (UniqueName: \"kubernetes.io/projected/d9e18e05-fb58-4fa9-b4be-be28657d52ee-kube-api-access-z85dd\") pod \"telemetry-operator-controller-manager-5b7d9bd964-gqwrj\" (UID: \"d9e18e05-fb58-4fa9-b4be-be28657d52ee\") " pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.844000 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.844024 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7kpp\" (UniqueName: \"kubernetes.io/projected/5d2ba7ed-1581-4cc2-957d-42ee5bf33781-kube-api-access-r7kpp\") pod \"test-operator-controller-manager-5f4f4847c9-gpsls\" (UID: \"5d2ba7ed-1581-4cc2-957d-42ee5bf33781\") " pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.845198 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.858355 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.881638 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7kpp\" (UniqueName: \"kubernetes.io/projected/5d2ba7ed-1581-4cc2-957d-42ee5bf33781-kube-api-access-r7kpp\") pod \"test-operator-controller-manager-5f4f4847c9-gpsls\" (UID: \"5d2ba7ed-1581-4cc2-957d-42ee5bf33781\") " pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.886342 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwb58\" (UniqueName: \"kubernetes.io/projected/bb6211ba-a514-4243-9204-0eb8954431cd-kube-api-access-nwb58\") pod \"swift-operator-controller-manager-7db4cd47d4-gmj7n\" (UID: \"bb6211ba-a514-4243-9204-0eb8954431cd\") " pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.891594 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z85dd\" (UniqueName: \"kubernetes.io/projected/d9e18e05-fb58-4fa9-b4be-be28657d52ee-kube-api-access-z85dd\") pod \"telemetry-operator-controller-manager-5b7d9bd964-gqwrj\" (UID: \"d9e18e05-fb58-4fa9-b4be-be28657d52ee\") " pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.897258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.917433 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.921500 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.922421 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.924622 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-2r22d" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.935206 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn"] Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.939280 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.945200 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzrb\" (UniqueName: \"kubernetes.io/projected/93fd4e69-37c4-485a-b780-6e366ebd0f65-kube-api-access-2qzrb\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.945268 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtvs4\" (UniqueName: \"kubernetes.io/projected/e355cfb0-5886-4e4a-a1cd-48b6e5b442f6-kube-api-access-mtvs4\") pod \"rabbitmq-cluster-operator-manager-79d8469568-zxtqn\" (UID: \"e355cfb0-5886-4e4a-a1cd-48b6e5b442f6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.945290 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.945854 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: E1004 11:03:48.945899 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert podName:93fd4e69-37c4-485a-b780-6e366ebd0f65 nodeName:}" failed. No retries permitted until 2025-10-04 11:03:49.445885501 +0000 UTC m=+806.738536390 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert") pod "openstack-operator-controller-manager-595579fdfd-9p69p" (UID: "93fd4e69-37c4-485a-b780-6e366ebd0f65") : secret "webhook-server-cert" not found Oct 04 11:03:48 crc kubenswrapper[4758]: I1004 11:03:48.958168 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.003978 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzrb\" (UniqueName: \"kubernetes.io/projected/93fd4e69-37c4-485a-b780-6e366ebd0f65-kube-api-access-2qzrb\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.046913 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtvs4\" (UniqueName: \"kubernetes.io/projected/e355cfb0-5886-4e4a-a1cd-48b6e5b442f6-kube-api-access-mtvs4\") pod \"rabbitmq-cluster-operator-manager-79d8469568-zxtqn\" (UID: \"e355cfb0-5886-4e4a-a1cd-48b6e5b442f6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.074169 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtvs4\" (UniqueName: \"kubernetes.io/projected/e355cfb0-5886-4e4a-a1cd-48b6e5b442f6-kube-api-access-mtvs4\") pod \"rabbitmq-cluster-operator-manager-79d8469568-zxtqn\" (UID: \"e355cfb0-5886-4e4a-a1cd-48b6e5b442f6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.149747 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b89709e-0a88-4e46-86e6-09e5b1b604da-cert\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.204175 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b89709e-0a88-4e46-86e6-09e5b1b604da-cert\") pod \"openstack-baremetal-operator-controller-manager-647c7ff67btjqdt\" (UID: \"5b89709e-0a88-4e46-86e6-09e5b1b604da\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.301558 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.329873 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.392446 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.392482 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.456691 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.463534 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:49 crc kubenswrapper[4758]: E1004 11:03:49.464332 4758 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 11:03:49 crc kubenswrapper[4758]: E1004 11:03:49.464406 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert podName:93fd4e69-37c4-485a-b780-6e366ebd0f65 nodeName:}" failed. No retries permitted until 2025-10-04 11:03:50.464384621 +0000 UTC m=+807.757035590 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert") pod "openstack-operator-controller-manager-595579fdfd-9p69p" (UID: "93fd4e69-37c4-485a-b780-6e366ebd0f65") : secret "webhook-server-cert" not found Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.632638 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr"] Oct 04 11:03:49 crc kubenswrapper[4758]: W1004 11:03:49.636555 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59448e96_d587_493a_b5c5_897aa1167152.slice/crio-6f67b06f586b12b6b1821816deb5533d0c4311dcfb67a270cde85d49c57339cf WatchSource:0}: Error finding container 6f67b06f586b12b6b1821816deb5533d0c4311dcfb67a270cde85d49c57339cf: Status 404 returned error can't find the container with id 6f67b06f586b12b6b1821816deb5533d0c4311dcfb67a270cde85d49c57339cf Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.767455 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.775481 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2bafb624-fe21-45cb-b787-192d4353378c-cert\") pod \"infra-operator-controller-manager-95c44c5bb-kg2lj\" (UID: \"2bafb624-fe21-45cb-b787-192d4353378c\") " pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.822520 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.830434 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.837384 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.848378 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" event={"ID":"aa921475-1667-4eff-8553-9a38928cf00f","Type":"ContainerStarted","Data":"d8eaab8c909291c82922aa6540133aaf8c85c91436a1cb6f81c1f448ff382812"} Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.860542 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.860649 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" event={"ID":"e261e6dd-0cdb-43b4-8981-5c34a7288fb3","Type":"ContainerStarted","Data":"8532ea991bfe54bb15f3b57e425f3f781a2a6b95dca55ec4767bbcf21d0dfa7b"} Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.866530 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls"] Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.873564 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" event={"ID":"59448e96-d587-493a-b5c5-897aa1167152","Type":"ContainerStarted","Data":"6f67b06f586b12b6b1821816deb5533d0c4311dcfb67a270cde85d49c57339cf"} Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.877061 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" event={"ID":"112ea09c-f473-4dbc-85d6-1e0ef7e6cf62","Type":"ContainerStarted","Data":"e87a8d9bef08ebdc1f450081294de09354886682c3b5c1b899babc0d9c624526"} Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.879376 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" event={"ID":"6c293821-0ef9-4bf3-9a2a-f188ac3b2151","Type":"ContainerStarted","Data":"e01dd4923fb7bc5b4adbcec8b03058b19b172bbdb1b93326cab7c0aa684bc4db"} Oct 04 11:03:49 crc kubenswrapper[4758]: I1004 11:03:49.881546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" event={"ID":"bfb96416-cb87-48e2-9ea7-71a899c3a620","Type":"ContainerStarted","Data":"7ff0a1d136bd0cd687de8a54ea8b96912352a1d86326f448557f362ce92e57ce"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.226907 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.250513 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls"] Oct 04 11:03:50 crc kubenswrapper[4758]: W1004 11:03:50.267780 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d2ba7ed_1581_4cc2_957d_42ee5bf33781.slice/crio-c3e64462500e545e0ae4cc13abb5dfa8200f8f805bb17d18825bd9d5fd140c97 WatchSource:0}: Error finding container c3e64462500e545e0ae4cc13abb5dfa8200f8f805bb17d18825bd9d5fd140c97: Status 404 returned error can't find the container with id c3e64462500e545e0ae4cc13abb5dfa8200f8f805bb17d18825bd9d5fd140c97 Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.281592 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.290399 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.327668 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77"] Oct 04 11:03:50 crc kubenswrapper[4758]: W1004 11:03:50.342410 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95aa14ab_a652_4eb4_97c1_7ab47431913b.slice/crio-64f71f11016a725b001086e3b3c3c23c92bc6090c4b9fec904342cff045b986f WatchSource:0}: Error finding container 64f71f11016a725b001086e3b3c3c23c92bc6090c4b9fec904342cff045b986f: Status 404 returned error can't find the container with id 64f71f11016a725b001086e3b3c3c23c92bc6090c4b9fec904342cff045b986f Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.355742 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx"] Oct 04 11:03:50 crc kubenswrapper[4758]: W1004 11:03:50.367615 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f16ce8c_ba77_4c4e_8f48_ee981b138a01.slice/crio-472215008d66af461dc1637929879d1157ec1ccfca80a020a583b30098c2b09b WatchSource:0}: Error finding container 472215008d66af461dc1637929879d1157ec1ccfca80a020a583b30098c2b09b: Status 404 returned error can't find the container with id 472215008d66af461dc1637929879d1157ec1ccfca80a020a583b30098c2b09b Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.368740 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:5d7bc83c30c70a32fbd727f0af07b1663bf96d7f80eaa000a3a157b58f065380,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nwb58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-7db4cd47d4-gmj7n_openstack-operators(bb6211ba-a514-4243-9204-0eb8954431cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.383549 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:a9aa51afd83a936d42ba7f6504eea549754d0b7e3bfec3af35039a3b92ee55c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4ccls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-74d4d6c8f6-7ndwj_openstack-operators(1f16ce8c-ba77-4c4e-8f48-ee981b138a01): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.384442 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a857f1d48b6bb38e77997baf7e7b7a9930aacab912b42ac10d1023d7e4e1dad0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dmq76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-647c7ff67btjqdt_openstack-operators(5b89709e-0a88-4e46-86e6-09e5b1b604da): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.384943 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:160f73b6bf799e057ba5b6df5ef91a77700fdae3c50d9df81398b1405b2b14f6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2rwzs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-768ddcd459-7zwcs_openstack-operators(4356f9f3-4e48-4ca8-92d4-f55914ded96b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.393984 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n"] Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.397139 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mtvs4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-zxtqn_openstack-operators(e355cfb0-5886-4e4a-a1cd-48b6e5b442f6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.398524 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" podUID="e355cfb0-5886-4e4a-a1cd-48b6e5b442f6" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.399706 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj"] Oct 04 11:03:50 crc kubenswrapper[4758]: W1004 11:03:50.419513 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bafb624_fe21_45cb_b787_192d4353378c.slice/crio-1ac429457c83b43865b8ed67be14ce6ece35efea98eaa98af2a6e4d1d1e51dae WatchSource:0}: Error finding container 1ac429457c83b43865b8ed67be14ce6ece35efea98eaa98af2a6e4d1d1e51dae: Status 404 returned error can't find the container with id 1ac429457c83b43865b8ed67be14ce6ece35efea98eaa98af2a6e4d1d1e51dae Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.420312 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.423755 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.439945 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.445915 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn"] Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.450657 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj"] Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.457181 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:b9006af5de38c6401e5d092c7671e94e6223f595675eaf251e3d45a1621ca032,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8qbnl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-95c44c5bb-kg2lj_openstack-operators(2bafb624-fe21-45cb-b787-192d4353378c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.484681 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.500529 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/93fd4e69-37c4-485a-b780-6e366ebd0f65-cert\") pod \"openstack-operator-controller-manager-595579fdfd-9p69p\" (UID: \"93fd4e69-37c4-485a-b780-6e366ebd0f65\") " pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.746657 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" podUID="1f16ce8c-ba77-4c4e-8f48-ee981b138a01" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.747711 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" podUID="5b89709e-0a88-4e46-86e6-09e5b1b604da" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.754797 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" podUID="2bafb624-fe21-45cb-b787-192d4353378c" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.759171 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" podUID="4356f9f3-4e48-4ca8-92d4-f55914ded96b" Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.762861 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" podUID="bb6211ba-a514-4243-9204-0eb8954431cd" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.788886 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.915441 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" event={"ID":"5b89709e-0a88-4e46-86e6-09e5b1b604da","Type":"ContainerStarted","Data":"19372edad80515c9d959f5af44bb792f6b3e414b2db694671311788fe1f6fb7d"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.915481 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" event={"ID":"5b89709e-0a88-4e46-86e6-09e5b1b604da","Type":"ContainerStarted","Data":"e112f459ee35f17037ae22862150e200ee9cfcaa7754689da5125aa81ee9ec13"} Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.916763 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a857f1d48b6bb38e77997baf7e7b7a9930aacab912b42ac10d1023d7e4e1dad0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" podUID="5b89709e-0a88-4e46-86e6-09e5b1b604da" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.928425 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" event={"ID":"d9e18e05-fb58-4fa9-b4be-be28657d52ee","Type":"ContainerStarted","Data":"3c5675d5443d318f796f27f59e98b94d641c3cc268924ef18ddb92d2c76d13bd"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.929970 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" event={"ID":"30c739ae-3a42-42c9-babb-03a8dd6b1a50","Type":"ContainerStarted","Data":"71c41d757bdc1e102fe5c2d3d8b434e78037f7a691dcde617384ce5331b234a7"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.932017 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" event={"ID":"dd449335-3579-4bea-a408-e8b92538f3fa","Type":"ContainerStarted","Data":"e818870901687c0c2c013fd97fbcbe16ac195a9e339c6be4809628ab7b9a8849"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.933153 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" event={"ID":"5d2ba7ed-1581-4cc2-957d-42ee5bf33781","Type":"ContainerStarted","Data":"c3e64462500e545e0ae4cc13abb5dfa8200f8f805bb17d18825bd9d5fd140c97"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.935488 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" event={"ID":"1f16ce8c-ba77-4c4e-8f48-ee981b138a01","Type":"ContainerStarted","Data":"b5ad1911b8ba28ba74ce934ea597e7b8684aaac980b597f6f39a1fc933e9eee2"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.935524 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" event={"ID":"1f16ce8c-ba77-4c4e-8f48-ee981b138a01","Type":"ContainerStarted","Data":"472215008d66af461dc1637929879d1157ec1ccfca80a020a583b30098c2b09b"} Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.937596 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:a9aa51afd83a936d42ba7f6504eea549754d0b7e3bfec3af35039a3b92ee55c8\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" podUID="1f16ce8c-ba77-4c4e-8f48-ee981b138a01" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.940437 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" event={"ID":"cc6e56f4-7374-4ffb-b5f2-55525602b161","Type":"ContainerStarted","Data":"bcd2f5636bda203b3506df26a6155911cd1bfbb43700443f4a0272cf74dcff4a"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.946447 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" event={"ID":"95aa14ab-a652-4eb4-97c1-7ab47431913b","Type":"ContainerStarted","Data":"64f71f11016a725b001086e3b3c3c23c92bc6090c4b9fec904342cff045b986f"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.968959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" event={"ID":"4356f9f3-4e48-4ca8-92d4-f55914ded96b","Type":"ContainerStarted","Data":"eb4789555d1121044e31c082f088f0b39551e3e8fb341fd2048af54f9f407f4d"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.969009 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" event={"ID":"4356f9f3-4e48-4ca8-92d4-f55914ded96b","Type":"ContainerStarted","Data":"eccf2c9ea0a4aa21052631ecd231a09ef2b617130dc64c58ca8cce0c6df5a266"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.971548 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" event={"ID":"f22d4cbc-37ae-4621-bc6d-4818642ecd29","Type":"ContainerStarted","Data":"8c81946c6b773eb51ad1344736eec213ff01a1e81d1eee6ba95b04263574d4c7"} Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.975724 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:160f73b6bf799e057ba5b6df5ef91a77700fdae3c50d9df81398b1405b2b14f6\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" podUID="4356f9f3-4e48-4ca8-92d4-f55914ded96b" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.979481 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" event={"ID":"e355cfb0-5886-4e4a-a1cd-48b6e5b442f6","Type":"ContainerStarted","Data":"0c3af92169b80df5188baa05f6c472c8066d55527b34fcba8a4b109622381aec"} Oct 04 11:03:50 crc kubenswrapper[4758]: E1004 11:03:50.985174 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" podUID="e355cfb0-5886-4e4a-a1cd-48b6e5b442f6" Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.991078 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" event={"ID":"7950908c-ddc0-434f-b28a-8d8687392a53","Type":"ContainerStarted","Data":"3c456ae6515cac54747432a0630e169612cd6dfe4d16168ae9d683e1b1c02a8c"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.992525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" event={"ID":"66ed67ad-7738-4701-9e08-0e45b5ac73d2","Type":"ContainerStarted","Data":"eb9bab836dc41bd5b2e9527ac9fe947a220c22a0f6faae55ba8012d1a0f8a6dc"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.998983 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" event={"ID":"2bafb624-fe21-45cb-b787-192d4353378c","Type":"ContainerStarted","Data":"5bac48706fb5ca7ef892210c9ce6f6d16f657415b16e483915660d787a6c8f4b"} Oct 04 11:03:50 crc kubenswrapper[4758]: I1004 11:03:50.999044 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" event={"ID":"2bafb624-fe21-45cb-b787-192d4353378c","Type":"ContainerStarted","Data":"1ac429457c83b43865b8ed67be14ce6ece35efea98eaa98af2a6e4d1d1e51dae"} Oct 04 11:03:51 crc kubenswrapper[4758]: E1004 11:03:51.007362 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b9006af5de38c6401e5d092c7671e94e6223f595675eaf251e3d45a1621ca032\\\"\"" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" podUID="2bafb624-fe21-45cb-b787-192d4353378c" Oct 04 11:03:51 crc kubenswrapper[4758]: I1004 11:03:51.036237 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" event={"ID":"bb6211ba-a514-4243-9204-0eb8954431cd","Type":"ContainerStarted","Data":"ea468675a25fe9293457ef8b8f37b8d348552b77a72ec4a822c270083aab78b2"} Oct 04 11:03:51 crc kubenswrapper[4758]: I1004 11:03:51.036282 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" event={"ID":"bb6211ba-a514-4243-9204-0eb8954431cd","Type":"ContainerStarted","Data":"d46ece84e24d58fb7ee732b80340d8f88c8936c07003e45d62bdf437bc8875ee"} Oct 04 11:03:51 crc kubenswrapper[4758]: E1004 11:03:51.046579 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:5d7bc83c30c70a32fbd727f0af07b1663bf96d7f80eaa000a3a157b58f065380\\\"\"" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" podUID="bb6211ba-a514-4243-9204-0eb8954431cd" Oct 04 11:03:51 crc kubenswrapper[4758]: I1004 11:03:51.416604 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p"] Oct 04 11:03:52 crc kubenswrapper[4758]: I1004 11:03:52.069070 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" event={"ID":"93fd4e69-37c4-485a-b780-6e366ebd0f65","Type":"ContainerStarted","Data":"3057854addeb2132b92abaf92d942cb05bddfe6ce179c5f656d5514ba52bb306"} Oct 04 11:03:52 crc kubenswrapper[4758]: I1004 11:03:52.070175 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" event={"ID":"93fd4e69-37c4-485a-b780-6e366ebd0f65","Type":"ContainerStarted","Data":"dade8edafde356b252257ae6f5519a999086b3f59cfa815da01e7a49e940d50f"} Oct 04 11:03:52 crc kubenswrapper[4758]: E1004 11:03:52.070825 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b9006af5de38c6401e5d092c7671e94e6223f595675eaf251e3d45a1621ca032\\\"\"" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" podUID="2bafb624-fe21-45cb-b787-192d4353378c" Oct 04 11:03:52 crc kubenswrapper[4758]: E1004 11:03:52.072304 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:5d7bc83c30c70a32fbd727f0af07b1663bf96d7f80eaa000a3a157b58f065380\\\"\"" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" podUID="bb6211ba-a514-4243-9204-0eb8954431cd" Oct 04 11:03:52 crc kubenswrapper[4758]: E1004 11:03:52.072800 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a857f1d48b6bb38e77997baf7e7b7a9930aacab912b42ac10d1023d7e4e1dad0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" podUID="5b89709e-0a88-4e46-86e6-09e5b1b604da" Oct 04 11:03:52 crc kubenswrapper[4758]: E1004 11:03:52.073283 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:160f73b6bf799e057ba5b6df5ef91a77700fdae3c50d9df81398b1405b2b14f6\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" podUID="4356f9f3-4e48-4ca8-92d4-f55914ded96b" Oct 04 11:03:52 crc kubenswrapper[4758]: E1004 11:03:52.076824 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:a9aa51afd83a936d42ba7f6504eea549754d0b7e3bfec3af35039a3b92ee55c8\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" podUID="1f16ce8c-ba77-4c4e-8f48-ee981b138a01" Oct 04 11:03:52 crc kubenswrapper[4758]: E1004 11:03:52.080839 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" podUID="e355cfb0-5886-4e4a-a1cd-48b6e5b442f6" Oct 04 11:03:53 crc kubenswrapper[4758]: I1004 11:03:53.102392 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" event={"ID":"93fd4e69-37c4-485a-b780-6e366ebd0f65","Type":"ContainerStarted","Data":"9d9a11c62162e87ba0f96c48d2109a5f2eac766bded03ec0f28183fe53ea332b"} Oct 04 11:03:53 crc kubenswrapper[4758]: I1004 11:03:53.102626 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:03:53 crc kubenswrapper[4758]: I1004 11:03:53.144644 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" podStartSLOduration=5.144626541 podStartE2EDuration="5.144626541s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:03:53.132216313 +0000 UTC m=+810.424867232" watchObservedRunningTime="2025-10-04 11:03:53.144626541 +0000 UTC m=+810.437277430" Oct 04 11:04:00 crc kubenswrapper[4758]: I1004 11:04:00.796928 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-595579fdfd-9p69p" Oct 04 11:04:01 crc kubenswrapper[4758]: I1004 11:04:01.249603 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:04:01 crc kubenswrapper[4758]: I1004 11:04:01.249978 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:04:01 crc kubenswrapper[4758]: I1004 11:04:01.250028 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:04:01 crc kubenswrapper[4758]: I1004 11:04:01.250653 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c234c874dc874e8dd12e5faf6f9512499bc90758f95287be1214207762ac29f"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:04:01 crc kubenswrapper[4758]: I1004 11:04:01.250712 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://5c234c874dc874e8dd12e5faf6f9512499bc90758f95287be1214207762ac29f" gracePeriod=600 Oct 04 11:04:02 crc kubenswrapper[4758]: I1004 11:04:02.181531 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="5c234c874dc874e8dd12e5faf6f9512499bc90758f95287be1214207762ac29f" exitCode=0 Oct 04 11:04:02 crc kubenswrapper[4758]: I1004 11:04:02.181720 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"5c234c874dc874e8dd12e5faf6f9512499bc90758f95287be1214207762ac29f"} Oct 04 11:04:02 crc kubenswrapper[4758]: I1004 11:04:02.181817 4758 scope.go:117] "RemoveContainer" containerID="4a4ba30925ab5365601b714b606e8b4bf9695f793e826fa68e869bdae0e9d640" Oct 04 11:04:04 crc kubenswrapper[4758]: E1004 11:04:04.101967 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:defecbd897a7c36829029a13b46219663f4162d603f5b824cee74ffa2428b2de" Oct 04 11:04:04 crc kubenswrapper[4758]: E1004 11:04:04.102672 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:defecbd897a7c36829029a13b46219663f4162d603f5b824cee74ffa2428b2de,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z85dd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5b7d9bd964-gqwrj_openstack-operators(d9e18e05-fb58-4fa9-b4be-be28657d52ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:04 crc kubenswrapper[4758]: E1004 11:04:04.515985 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:8f7225fc86ea0d0950b5370a9f9089d65c2b324be871987521d97bff1583ae3e" Oct 04 11:04:04 crc kubenswrapper[4758]: E1004 11:04:04.517269 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:8f7225fc86ea0d0950b5370a9f9089d65c2b324be871987521d97bff1583ae3e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-54q5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b97cccc57-g8m77_openstack-operators(dd449335-3579-4bea-a408-e8b92538f3fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:04 crc kubenswrapper[4758]: E1004 11:04:04.912045 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:0c7adc09b0c90a1503823cafc53e83086742a9f660fad42a76407c852301b644" Oct 04 11:04:04 crc kubenswrapper[4758]: E1004 11:04:04.912241 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:0c7adc09b0c90a1503823cafc53e83086742a9f660fad42a76407c852301b644,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzzx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-57b88b6ccd-m24qr_openstack-operators(66ed67ad-7738-4701-9e08-0e45b5ac73d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:06 crc kubenswrapper[4758]: E1004 11:04:06.097546 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:1759091cc372becb5f1f328d93a38bcf88ebb0a71a24ada3a6a728a924290b6e" Oct 04 11:04:06 crc kubenswrapper[4758]: E1004 11:04:06.098168 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:1759091cc372becb5f1f328d93a38bcf88ebb0a71a24ada3a6a728a924290b6e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v9gts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-746ff9ccdc-d269t_openstack-operators(7950908c-ddc0-434f-b28a-8d8687392a53): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:06 crc kubenswrapper[4758]: E1004 11:04:06.584777 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:01837b37bdb95d24e59c77e72330145a995c123a24d86f548e641a9d9e53293c" Oct 04 11:04:06 crc kubenswrapper[4758]: E1004 11:04:06.584968 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:01837b37bdb95d24e59c77e72330145a995c123a24d86f548e641a9d9e53293c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-clr4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-578455559c-l7m4w_openstack-operators(bfb96416-cb87-48e2-9ea7-71a899c3a620): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:08 crc kubenswrapper[4758]: E1004 11:04:08.237987 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:71f282fa63dc1cf9d02db6a527d65d55ab1cbbd36d91ee194185bc14089b2513" Oct 04 11:04:08 crc kubenswrapper[4758]: E1004 11:04:08.238172 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:71f282fa63dc1cf9d02db6a527d65d55ab1cbbd36d91ee194185bc14089b2513,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ppxs7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-dd56696bb-dm2ls_openstack-operators(f22d4cbc-37ae-4621-bc6d-4818642ecd29): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:08 crc kubenswrapper[4758]: E1004 11:04:08.643946 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:4421d570ffdc5299acd303071d9256dfb83a43f127ace2c68b6465b417dd731e" Oct 04 11:04:08 crc kubenswrapper[4758]: E1004 11:04:08.644116 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:4421d570ffdc5299acd303071d9256dfb83a43f127ace2c68b6465b417dd731e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lj9h5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-697cc69689-gmq9x_openstack-operators(30c739ae-3a42-42c9-babb-03a8dd6b1a50): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:09 crc kubenswrapper[4758]: E1004 11:04:09.975058 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:759b6819170324524e39ee25d0fe171e6a9a638e62904944a596cfab42481ef7" Oct 04 11:04:09 crc kubenswrapper[4758]: E1004 11:04:09.975727 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:759b6819170324524e39ee25d0fe171e6a9a638e62904944a596cfab42481ef7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r7kpp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5f4f4847c9-gpsls_openstack-operators(5d2ba7ed-1581-4cc2-957d-42ee5bf33781): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.695756 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d2vpw"] Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.697538 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.715861 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2vpw"] Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.844744 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7gzz\" (UniqueName: \"kubernetes.io/projected/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-kube-api-access-v7gzz\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.844799 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-catalog-content\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.844955 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-utilities\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.946620 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-utilities\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.946943 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-catalog-content\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.946960 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7gzz\" (UniqueName: \"kubernetes.io/projected/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-kube-api-access-v7gzz\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.947249 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-utilities\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.947482 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-catalog-content\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:13 crc kubenswrapper[4758]: I1004 11:04:13.967686 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7gzz\" (UniqueName: \"kubernetes.io/projected/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-kube-api-access-v7gzz\") pod \"redhat-operators-d2vpw\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:14 crc kubenswrapper[4758]: I1004 11:04:14.027251 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:20 crc kubenswrapper[4758]: E1004 11:04:20.793544 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a857f1d48b6bb38e77997baf7e7b7a9930aacab912b42ac10d1023d7e4e1dad0" Oct 04 11:04:20 crc kubenswrapper[4758]: E1004 11:04:20.794466 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a857f1d48b6bb38e77997baf7e7b7a9930aacab912b42ac10d1023d7e4e1dad0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dmq76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-647c7ff67btjqdt_openstack-operators(5b89709e-0a88-4e46-86e6-09e5b1b604da): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:20 crc kubenswrapper[4758]: E1004 11:04:20.795818 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" podUID="5b89709e-0a88-4e46-86e6-09e5b1b604da" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.157580 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.157749 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mtvs4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-79d8469568-zxtqn_openstack-operators(e355cfb0-5886-4e4a-a1cd-48b6e5b442f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.159024 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" podUID="e355cfb0-5886-4e4a-a1cd-48b6e5b442f6" Oct 04 11:04:21 crc kubenswrapper[4758]: I1004 11:04:21.632075 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d2vpw"] Oct 04 11:04:21 crc kubenswrapper[4758]: W1004 11:04:21.648450 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d3122ac_f3b9_49c3_96fa_58109aa83dc6.slice/crio-9addf9f909b43f5bcfe85919959db029d6f084c5ecc040244c0f1144b2be9d9c WatchSource:0}: Error finding container 9addf9f909b43f5bcfe85919959db029d6f084c5ecc040244c0f1144b2be9d9c: Status 404 returned error can't find the container with id 9addf9f909b43f5bcfe85919959db029d6f084c5ecc040244c0f1144b2be9d9c Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.755081 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" podUID="d9e18e05-fb58-4fa9-b4be-be28657d52ee" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.755144 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" podUID="bfb96416-cb87-48e2-9ea7-71a899c3a620" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.772656 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" podUID="66ed67ad-7738-4701-9e08-0e45b5ac73d2" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.772705 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" podUID="7950908c-ddc0-434f-b28a-8d8687392a53" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.774272 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" podUID="dd449335-3579-4bea-a408-e8b92538f3fa" Oct 04 11:04:21 crc kubenswrapper[4758]: E1004 11:04:21.854310 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" podUID="30c739ae-3a42-42c9-babb-03a8dd6b1a50" Oct 04 11:04:22 crc kubenswrapper[4758]: E1004 11:04:22.039374 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" podUID="f22d4cbc-37ae-4621-bc6d-4818642ecd29" Oct 04 11:04:22 crc kubenswrapper[4758]: E1004 11:04:22.045133 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" podUID="5d2ba7ed-1581-4cc2-957d-42ee5bf33781" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.323318 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" event={"ID":"112ea09c-f473-4dbc-85d6-1e0ef7e6cf62","Type":"ContainerStarted","Data":"57a8d3be8af962a30679d91c6fccf01c7b66ecfeb47f550b4cec8cbc614a1c51"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.324403 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" event={"ID":"95aa14ab-a652-4eb4-97c1-7ab47431913b","Type":"ContainerStarted","Data":"546536df633b23dd8ce594a5b4e97f8479597191afd04d920bb45ccf22f9944a"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.325841 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" event={"ID":"f22d4cbc-37ae-4621-bc6d-4818642ecd29","Type":"ContainerStarted","Data":"4b955ba0914d614e0f5f31ba8cabf4d290b8e894ac6d0259b792acf0acd9bc35"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.328268 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" event={"ID":"bfb96416-cb87-48e2-9ea7-71a899c3a620","Type":"ContainerStarted","Data":"4efb8cf971866a2224037fa1da815da2c2f48d82484a515be6c5b5e374d0b916"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.336002 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" event={"ID":"66ed67ad-7738-4701-9e08-0e45b5ac73d2","Type":"ContainerStarted","Data":"176b54cf0763f26c43de56909320c027b36396e6844b1a41b5243ace79c928f0"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.352438 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" event={"ID":"4356f9f3-4e48-4ca8-92d4-f55914ded96b","Type":"ContainerStarted","Data":"d3d4b6e5e0a23d6af62645a4cf11482df244bbb6e81a5ffd75b4db6ae0b2d0cc"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.353033 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.372607 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" event={"ID":"bb6211ba-a514-4243-9204-0eb8954431cd","Type":"ContainerStarted","Data":"465af0c7624821e6311906703324aaf490f1938cde03cbed09361edc35d4e42d"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.372955 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.376707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" event={"ID":"7950908c-ddc0-434f-b28a-8d8687392a53","Type":"ContainerStarted","Data":"e014646f649a8192a950307ac12dfdb715046f257381a0f6abe2a9ac6ac92ec5"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.380730 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" event={"ID":"59448e96-d587-493a-b5c5-897aa1167152","Type":"ContainerStarted","Data":"f4f975e5fb48ba21af69b2069527b1b4e63a1a636fab5e5f0518e13d8d84a90e"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.388204 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" event={"ID":"30c739ae-3a42-42c9-babb-03a8dd6b1a50","Type":"ContainerStarted","Data":"7a4f512f1e7155e0b2bbccbe59b23018c593a088e17254c3ac033111e4e281a9"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.393712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" event={"ID":"1f16ce8c-ba77-4c4e-8f48-ee981b138a01","Type":"ContainerStarted","Data":"aad0a6b25e3c1d576656e5c5be5c85c98d8f0706369758915475cbeaa3c3fbb5"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.395436 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.415304 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"2b150cc50b08a3a5c33e13466357eafa5f32ab57cc3437792ce3dec5314cb49d"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.420038 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" event={"ID":"2bafb624-fe21-45cb-b787-192d4353378c","Type":"ContainerStarted","Data":"e90ef0162a7d0cd0f7809e29f8ccb7e1796b614c721fda6c25b45d5ab1ab3836"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.420683 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.428338 4758 generic.go:334] "Generic (PLEG): container finished" podID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerID="e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357" exitCode=0 Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.428419 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerDied","Data":"e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.428460 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerStarted","Data":"9addf9f909b43f5bcfe85919959db029d6f084c5ecc040244c0f1144b2be9d9c"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.442190 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" event={"ID":"d9e18e05-fb58-4fa9-b4be-be28657d52ee","Type":"ContainerStarted","Data":"f4fe1763a9628e75fe6074ccd5b6b1493d180398d646ae90a926de984317d8ec"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.455346 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" event={"ID":"5d2ba7ed-1581-4cc2-957d-42ee5bf33781","Type":"ContainerStarted","Data":"2e26922a11eb7fa78208534e5ebd8730222ddb3b1c8b326f73861482e14f9371"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.473246 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" event={"ID":"aa921475-1667-4eff-8553-9a38928cf00f","Type":"ContainerStarted","Data":"2292ea40538dab7e3dc41c3aec10c6dbce173f516d25a5b520b753fbee484912"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.473289 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" event={"ID":"aa921475-1667-4eff-8553-9a38928cf00f","Type":"ContainerStarted","Data":"707512f914a8db5bf88efcde00f44936211a6bfe3c3c5fd95d08c7282d462fc6"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.473965 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.495333 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" event={"ID":"6c293821-0ef9-4bf3-9a2a-f188ac3b2151","Type":"ContainerStarted","Data":"be29ff0b5c92bd21d31e10274237843a02f2fb51c73a76d37633d2c8bca3c086"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.495391 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" event={"ID":"6c293821-0ef9-4bf3-9a2a-f188ac3b2151","Type":"ContainerStarted","Data":"9b25e932e78bf1161e1635f3e10341dffd5f140dbadd32f1c7f5b36d02d0b473"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.496072 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.523354 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" event={"ID":"e261e6dd-0cdb-43b4-8981-5c34a7288fb3","Type":"ContainerStarted","Data":"b19e889e2ab3a4850b2cc99b52184181607a271212196bf7c86f838b4348af5b"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.530338 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" event={"ID":"dd449335-3579-4bea-a408-e8b92538f3fa","Type":"ContainerStarted","Data":"499c546555b9d1f9ba8829f479288e135fa1947881d2e4bea1e5b8ecabd1c9b8"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.544176 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" event={"ID":"cc6e56f4-7374-4ffb-b5f2-55525602b161","Type":"ContainerStarted","Data":"7e846c4a328f777160309d2f39fe0d4f719f356bbd5b2429aad26bc5cda04146"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.544480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" event={"ID":"cc6e56f4-7374-4ffb-b5f2-55525602b161","Type":"ContainerStarted","Data":"bd25b5e138ad9f292ce7fdb351150b28da91941f13ff0f21541d89f652061539"} Oct 04 11:04:22 crc kubenswrapper[4758]: I1004 11:04:22.545154 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.490754 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" podStartSLOduration=13.453142738 podStartE2EDuration="35.490738586s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.828357721 +0000 UTC m=+807.121008610" lastFinishedPulling="2025-10-04 11:04:11.865953569 +0000 UTC m=+829.158604458" observedRunningTime="2025-10-04 11:04:23.428244824 +0000 UTC m=+840.720895713" watchObservedRunningTime="2025-10-04 11:04:23.490738586 +0000 UTC m=+840.783389475" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.536455 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" podStartSLOduration=15.53377999 podStartE2EDuration="36.536440296s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.444309139 +0000 UTC m=+806.736960028" lastFinishedPulling="2025-10-04 11:04:10.446969445 +0000 UTC m=+827.739620334" observedRunningTime="2025-10-04 11:04:23.531320133 +0000 UTC m=+840.823971022" watchObservedRunningTime="2025-10-04 11:04:23.536440296 +0000 UTC m=+840.829091175" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.558056 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" podStartSLOduration=5.813511312 podStartE2EDuration="36.558038611s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.457024398 +0000 UTC m=+807.749675287" lastFinishedPulling="2025-10-04 11:04:21.201551687 +0000 UTC m=+838.494202586" observedRunningTime="2025-10-04 11:04:23.552614059 +0000 UTC m=+840.845264948" watchObservedRunningTime="2025-10-04 11:04:23.558038611 +0000 UTC m=+840.850689500" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.567013 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" event={"ID":"95aa14ab-a652-4eb4-97c1-7ab47431913b","Type":"ContainerStarted","Data":"f0367c1e7417d562e50a7b13322b7573996f8a7f692f01f2f0fb9098be43be4d"} Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.567828 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.579236 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" event={"ID":"e261e6dd-0cdb-43b4-8981-5c34a7288fb3","Type":"ContainerStarted","Data":"9c630cc60756e3a8b90889a8e73f9415e48d97b6b21ba09e4a198092fceaba94"} Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.579459 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.588096 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" podStartSLOduration=15.524876352 podStartE2EDuration="35.588083253s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.382751456 +0000 UTC m=+807.675402345" lastFinishedPulling="2025-10-04 11:04:10.445958357 +0000 UTC m=+827.738609246" observedRunningTime="2025-10-04 11:04:23.584365499 +0000 UTC m=+840.877016388" watchObservedRunningTime="2025-10-04 11:04:23.588083253 +0000 UTC m=+840.880734142" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.593617 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" event={"ID":"59448e96-d587-493a-b5c5-897aa1167152","Type":"ContainerStarted","Data":"802c6247aa3ad7cb1d2bdb1ebea669ad2011532c895545701e9ff8086b55fde4"} Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.594334 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.596906 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" event={"ID":"112ea09c-f473-4dbc-85d6-1e0ef7e6cf62","Type":"ContainerStarted","Data":"bddeb233a459fe354256d2806b5d5a832d33e473973107c9d256d7ed923dfa96"} Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.596932 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.640310 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" podStartSLOduration=5.234702484 podStartE2EDuration="35.640292846s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.383059805 +0000 UTC m=+807.675710694" lastFinishedPulling="2025-10-04 11:04:20.788650147 +0000 UTC m=+838.081301056" observedRunningTime="2025-10-04 11:04:23.637041395 +0000 UTC m=+840.929692284" watchObservedRunningTime="2025-10-04 11:04:23.640292846 +0000 UTC m=+840.932943736" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.700929 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" podStartSLOduration=4.81488828 podStartE2EDuration="35.700910905s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.367724985 +0000 UTC m=+807.660375874" lastFinishedPulling="2025-10-04 11:04:21.2537476 +0000 UTC m=+838.546398499" observedRunningTime="2025-10-04 11:04:23.675984027 +0000 UTC m=+840.968634916" watchObservedRunningTime="2025-10-04 11:04:23.700910905 +0000 UTC m=+840.993561794" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.758618 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" podStartSLOduration=4.926110108 podStartE2EDuration="35.758600522s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.383416725 +0000 UTC m=+807.676067614" lastFinishedPulling="2025-10-04 11:04:21.215907139 +0000 UTC m=+838.508558028" observedRunningTime="2025-10-04 11:04:23.751630527 +0000 UTC m=+841.044281416" watchObservedRunningTime="2025-10-04 11:04:23.758600522 +0000 UTC m=+841.051251411" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.814603 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" podStartSLOduration=14.296060118 podStartE2EDuration="35.814583531s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.346705006 +0000 UTC m=+807.639355895" lastFinishedPulling="2025-10-04 11:04:11.865228419 +0000 UTC m=+829.157879308" observedRunningTime="2025-10-04 11:04:23.797941244 +0000 UTC m=+841.090592133" watchObservedRunningTime="2025-10-04 11:04:23.814583531 +0000 UTC m=+841.107234420" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.850000 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" podStartSLOduration=14.369324117 podStartE2EDuration="36.849975792s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.376179929 +0000 UTC m=+806.668830818" lastFinishedPulling="2025-10-04 11:04:11.856831604 +0000 UTC m=+829.149482493" observedRunningTime="2025-10-04 11:04:23.82528023 +0000 UTC m=+841.117931119" watchObservedRunningTime="2025-10-04 11:04:23.849975792 +0000 UTC m=+841.142626681" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.935965 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" podStartSLOduration=14.528121129 podStartE2EDuration="36.935921861s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.449803003 +0000 UTC m=+806.742453892" lastFinishedPulling="2025-10-04 11:04:11.857603735 +0000 UTC m=+829.150254624" observedRunningTime="2025-10-04 11:04:23.858524802 +0000 UTC m=+841.151175691" watchObservedRunningTime="2025-10-04 11:04:23.935921861 +0000 UTC m=+841.228572750" Oct 04 11:04:23 crc kubenswrapper[4758]: I1004 11:04:23.970657 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" podStartSLOduration=14.752057434 podStartE2EDuration="36.970643004s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.639012715 +0000 UTC m=+806.931663604" lastFinishedPulling="2025-10-04 11:04:11.857598295 +0000 UTC m=+829.150249174" observedRunningTime="2025-10-04 11:04:23.901234139 +0000 UTC m=+841.193885038" watchObservedRunningTime="2025-10-04 11:04:23.970643004 +0000 UTC m=+841.263293893" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.605128 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" event={"ID":"f22d4cbc-37ae-4621-bc6d-4818642ecd29","Type":"ContainerStarted","Data":"888dcef011220f7e59487a88a8ce8ab93657aca2f92995cad54e50d90407949d"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.606567 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.608707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" event={"ID":"d9e18e05-fb58-4fa9-b4be-be28657d52ee","Type":"ContainerStarted","Data":"ddd60253bc4f077e52cc2d4754b58de337e01424a017d9d565a3213d4a7cfbfa"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.609350 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.611733 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" event={"ID":"dd449335-3579-4bea-a408-e8b92538f3fa","Type":"ContainerStarted","Data":"c23f22e7e9b554145ffaeb6f1c6b8f9219630b4cdb8fa6f184af098819126903"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.612373 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.613972 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" event={"ID":"66ed67ad-7738-4701-9e08-0e45b5ac73d2","Type":"ContainerStarted","Data":"30a611da885784a9c10b70b094e9e805dfeee6b50ae3e4fe547f935f39e0fec9"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.614045 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.616122 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerStarted","Data":"211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.619929 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" event={"ID":"7950908c-ddc0-434f-b28a-8d8687392a53","Type":"ContainerStarted","Data":"d051d3134a04ad56fb6ea3c92633fe48900e8ccfbb246f0e0ae88a3f830d16e6"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.620160 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.622205 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" event={"ID":"5d2ba7ed-1581-4cc2-957d-42ee5bf33781","Type":"ContainerStarted","Data":"37f61de987905c31dfeffa793c8fb612121d38eb769a712fb1c2d57a3f542150"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.622354 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.626444 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" event={"ID":"30c739ae-3a42-42c9-babb-03a8dd6b1a50","Type":"ContainerStarted","Data":"888ed9cf2856fe11d7a202f856d965795bb7b3b03897e809f988fb264395a1ce"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.626544 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.631140 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" event={"ID":"bfb96416-cb87-48e2-9ea7-71a899c3a620","Type":"ContainerStarted","Data":"2f41839ce887febd48de83fff34c51b378c0fb90a833ba1a274d2a032ee6f967"} Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.643488 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" podStartSLOduration=3.743991459 podStartE2EDuration="37.643460658s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.87614388 +0000 UTC m=+807.168794769" lastFinishedPulling="2025-10-04 11:04:23.775613079 +0000 UTC m=+841.068263968" observedRunningTime="2025-10-04 11:04:24.641112582 +0000 UTC m=+841.933763471" watchObservedRunningTime="2025-10-04 11:04:24.643460658 +0000 UTC m=+841.936111537" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.689092 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" podStartSLOduration=3.791409598 podStartE2EDuration="37.689074906s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.876647964 +0000 UTC m=+807.169298853" lastFinishedPulling="2025-10-04 11:04:23.774313272 +0000 UTC m=+841.066964161" observedRunningTime="2025-10-04 11:04:24.687126302 +0000 UTC m=+841.979777181" watchObservedRunningTime="2025-10-04 11:04:24.689074906 +0000 UTC m=+841.981725795" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.729137 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" podStartSLOduration=3.598399552 podStartE2EDuration="36.729122448s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.332425226 +0000 UTC m=+807.625076115" lastFinishedPulling="2025-10-04 11:04:23.463148122 +0000 UTC m=+840.755799011" observedRunningTime="2025-10-04 11:04:24.728781799 +0000 UTC m=+842.021432698" watchObservedRunningTime="2025-10-04 11:04:24.729122448 +0000 UTC m=+842.021773347" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.801757 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" podStartSLOduration=3.591017177 podStartE2EDuration="36.801740435s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.239021129 +0000 UTC m=+807.531672018" lastFinishedPulling="2025-10-04 11:04:23.449744387 +0000 UTC m=+840.742395276" observedRunningTime="2025-10-04 11:04:24.800356227 +0000 UTC m=+842.093007116" watchObservedRunningTime="2025-10-04 11:04:24.801740435 +0000 UTC m=+842.094391324" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.857548 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" podStartSLOduration=3.361859855 podStartE2EDuration="36.857531288s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.355261996 +0000 UTC m=+807.647912885" lastFinishedPulling="2025-10-04 11:04:23.850933429 +0000 UTC m=+841.143584318" observedRunningTime="2025-10-04 11:04:24.853134539 +0000 UTC m=+842.145785438" watchObservedRunningTime="2025-10-04 11:04:24.857531288 +0000 UTC m=+842.150182167" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.908017 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" podStartSLOduration=4.036580805 podStartE2EDuration="37.908000168s" podCreationTimestamp="2025-10-04 11:03:47 +0000 UTC" firstStartedPulling="2025-10-04 11:03:49.838935967 +0000 UTC m=+807.131586856" lastFinishedPulling="2025-10-04 11:04:23.71035533 +0000 UTC m=+841.003006219" observedRunningTime="2025-10-04 11:04:24.887599558 +0000 UTC m=+842.180250447" watchObservedRunningTime="2025-10-04 11:04:24.908000168 +0000 UTC m=+842.200651057" Oct 04 11:04:24 crc kubenswrapper[4758]: I1004 11:04:24.910314 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" podStartSLOduration=3.697792701 podStartE2EDuration="36.91030448s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.288544396 +0000 UTC m=+807.581195285" lastFinishedPulling="2025-10-04 11:04:23.501056175 +0000 UTC m=+840.793707064" observedRunningTime="2025-10-04 11:04:24.905242183 +0000 UTC m=+842.197893072" watchObservedRunningTime="2025-10-04 11:04:24.91030448 +0000 UTC m=+842.202955359" Oct 04 11:04:25 crc kubenswrapper[4758]: I1004 11:04:25.640483 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:04:26 crc kubenswrapper[4758]: I1004 11:04:26.648891 4758 generic.go:334] "Generic (PLEG): container finished" podID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerID="211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c" exitCode=0 Oct 04 11:04:26 crc kubenswrapper[4758]: I1004 11:04:26.649847 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerDied","Data":"211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c"} Oct 04 11:04:26 crc kubenswrapper[4758]: I1004 11:04:26.694040 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" podStartSLOduration=5.279640222 podStartE2EDuration="38.694013237s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.361711927 +0000 UTC m=+807.654362816" lastFinishedPulling="2025-10-04 11:04:23.776084942 +0000 UTC m=+841.068735831" observedRunningTime="2025-10-04 11:04:24.939987849 +0000 UTC m=+842.232638738" watchObservedRunningTime="2025-10-04 11:04:26.694013237 +0000 UTC m=+843.986664166" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.182744 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d49d7896f-mgphn" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.202716 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-866c4585d9-frqpq" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.212413 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-5d8bb764f9-77mnr" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.275719 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6b95d97c7f-lbtmk" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.438711 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-795ccf785c-jtcj2" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.555625 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6679c59c97-r7gpj" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.583755 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-74d4d6c8f6-7ndwj" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.668276 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerStarted","Data":"fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2"} Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.683709 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d2vpw" podStartSLOduration=10.391550321 podStartE2EDuration="15.683695025s" podCreationTimestamp="2025-10-04 11:04:13 +0000 UTC" firstStartedPulling="2025-10-04 11:04:22.432378808 +0000 UTC m=+839.725029697" lastFinishedPulling="2025-10-04 11:04:27.724523512 +0000 UTC m=+845.017174401" observedRunningTime="2025-10-04 11:04:28.68126886 +0000 UTC m=+845.973919749" watchObservedRunningTime="2025-10-04 11:04:28.683695025 +0000 UTC m=+845.976345904" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.767790 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57b88b6ccd-m24qr" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.812514 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-768ddcd459-7zwcs" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.860909 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d879db64d-r49wx" Oct 04 11:04:28 crc kubenswrapper[4758]: I1004 11:04:28.942683 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7db4cd47d4-gmj7n" Oct 04 11:04:29 crc kubenswrapper[4758]: I1004 11:04:29.843404 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-95c44c5bb-kg2lj" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.131951 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-24p8j"] Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.134204 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.161117 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24p8j"] Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.180212 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2566\" (UniqueName: \"kubernetes.io/projected/0ff99867-758e-4dcc-a7eb-74fe410cf52d-kube-api-access-d2566\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.180284 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-utilities\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.180309 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-catalog-content\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.281259 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2566\" (UniqueName: \"kubernetes.io/projected/0ff99867-758e-4dcc-a7eb-74fe410cf52d-kube-api-access-d2566\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.281735 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-utilities\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.281804 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-catalog-content\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.282331 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-catalog-content\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.282596 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-utilities\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.314980 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2566\" (UniqueName: \"kubernetes.io/projected/0ff99867-758e-4dcc-a7eb-74fe410cf52d-kube-api-access-d2566\") pod \"community-operators-24p8j\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:32 crc kubenswrapper[4758]: E1004 11:04:32.328015 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a857f1d48b6bb38e77997baf7e7b7a9930aacab912b42ac10d1023d7e4e1dad0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" podUID="5b89709e-0a88-4e46-86e6-09e5b1b604da" Oct 04 11:04:32 crc kubenswrapper[4758]: I1004 11:04:32.448710 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:33 crc kubenswrapper[4758]: I1004 11:04:33.157709 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24p8j"] Oct 04 11:04:33 crc kubenswrapper[4758]: W1004 11:04:33.161342 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ff99867_758e_4dcc_a7eb_74fe410cf52d.slice/crio-7efb3a1785c037c91a7c7215f9e38545e99b8975b3b4a79427c2a91bb6b5891c WatchSource:0}: Error finding container 7efb3a1785c037c91a7c7215f9e38545e99b8975b3b4a79427c2a91bb6b5891c: Status 404 returned error can't find the container with id 7efb3a1785c037c91a7c7215f9e38545e99b8975b3b4a79427c2a91bb6b5891c Oct 04 11:04:33 crc kubenswrapper[4758]: I1004 11:04:33.711046 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerStarted","Data":"114123af40c52f01ec336dc7fe1b0c0e1f45880e50690ef81506a1494c381a26"} Oct 04 11:04:33 crc kubenswrapper[4758]: I1004 11:04:33.711468 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerStarted","Data":"7efb3a1785c037c91a7c7215f9e38545e99b8975b3b4a79427c2a91bb6b5891c"} Oct 04 11:04:34 crc kubenswrapper[4758]: I1004 11:04:34.028274 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:34 crc kubenswrapper[4758]: I1004 11:04:34.028352 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:04:34 crc kubenswrapper[4758]: I1004 11:04:34.718979 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerDied","Data":"114123af40c52f01ec336dc7fe1b0c0e1f45880e50690ef81506a1494c381a26"} Oct 04 11:04:34 crc kubenswrapper[4758]: I1004 11:04:34.718922 4758 generic.go:334] "Generic (PLEG): container finished" podID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerID="114123af40c52f01ec336dc7fe1b0c0e1f45880e50690ef81506a1494c381a26" exitCode=0 Oct 04 11:04:35 crc kubenswrapper[4758]: I1004 11:04:35.083333 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d2vpw" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" probeResult="failure" output=< Oct 04 11:04:35 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:04:35 crc kubenswrapper[4758]: > Oct 04 11:04:35 crc kubenswrapper[4758]: I1004 11:04:35.727402 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerStarted","Data":"48ebb121aea7cb2bba239a1f5b9ccb954b7d8577fd04f8b0656ab5098b2e27db"} Oct 04 11:04:36 crc kubenswrapper[4758]: E1004 11:04:36.327044 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:225524223bf2a7f3a4ce95958fc9ca6fdab02745fb70374e8ff5bf1ddaceda4b\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" podUID="e355cfb0-5886-4e4a-a1cd-48b6e5b442f6" Oct 04 11:04:36 crc kubenswrapper[4758]: I1004 11:04:36.736075 4758 generic.go:334] "Generic (PLEG): container finished" podID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerID="48ebb121aea7cb2bba239a1f5b9ccb954b7d8577fd04f8b0656ab5098b2e27db" exitCode=0 Oct 04 11:04:36 crc kubenswrapper[4758]: I1004 11:04:36.736152 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerDied","Data":"48ebb121aea7cb2bba239a1f5b9ccb954b7d8577fd04f8b0656ab5098b2e27db"} Oct 04 11:04:37 crc kubenswrapper[4758]: I1004 11:04:37.744501 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerStarted","Data":"71c46a8a6a1a7cbf5c517f918dd5386a874f02b70e2b923c6a0744928ee27e92"} Oct 04 11:04:37 crc kubenswrapper[4758]: I1004 11:04:37.771073 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-24p8j" podStartSLOduration=3.35956926 podStartE2EDuration="5.771040962s" podCreationTimestamp="2025-10-04 11:04:32 +0000 UTC" firstStartedPulling="2025-10-04 11:04:34.720758519 +0000 UTC m=+852.013409398" lastFinishedPulling="2025-10-04 11:04:37.132230201 +0000 UTC m=+854.424881100" observedRunningTime="2025-10-04 11:04:37.766176671 +0000 UTC m=+855.058827580" watchObservedRunningTime="2025-10-04 11:04:37.771040962 +0000 UTC m=+855.063691891" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.291579 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-697cc69689-gmq9x" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.311303 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-dd56696bb-dm2ls" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.351433 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-578455559c-l7m4w" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.674656 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-746ff9ccdc-d269t" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.899854 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b97cccc57-g8m77" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.921219 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5f4f4847c9-gpsls" Oct 04 11:04:38 crc kubenswrapper[4758]: I1004 11:04:38.962652 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5b7d9bd964-gqwrj" Oct 04 11:04:42 crc kubenswrapper[4758]: I1004 11:04:42.450341 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:42 crc kubenswrapper[4758]: I1004 11:04:42.451068 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:42 crc kubenswrapper[4758]: I1004 11:04:42.533379 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:42 crc kubenswrapper[4758]: I1004 11:04:42.876801 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:42 crc kubenswrapper[4758]: I1004 11:04:42.952299 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24p8j"] Oct 04 11:04:44 crc kubenswrapper[4758]: I1004 11:04:44.794979 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-24p8j" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="registry-server" containerID="cri-o://71c46a8a6a1a7cbf5c517f918dd5386a874f02b70e2b923c6a0744928ee27e92" gracePeriod=2 Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.064417 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d2vpw" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" probeResult="failure" output=< Oct 04 11:04:45 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:04:45 crc kubenswrapper[4758]: > Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.809810 4758 generic.go:334] "Generic (PLEG): container finished" podID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerID="71c46a8a6a1a7cbf5c517f918dd5386a874f02b70e2b923c6a0744928ee27e92" exitCode=0 Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.809979 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerDied","Data":"71c46a8a6a1a7cbf5c517f918dd5386a874f02b70e2b923c6a0744928ee27e92"} Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.927152 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.972418 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-catalog-content\") pod \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.972494 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-utilities\") pod \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.972575 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2566\" (UniqueName: \"kubernetes.io/projected/0ff99867-758e-4dcc-a7eb-74fe410cf52d-kube-api-access-d2566\") pod \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\" (UID: \"0ff99867-758e-4dcc-a7eb-74fe410cf52d\") " Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.974643 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-utilities" (OuterVolumeSpecName: "utilities") pod "0ff99867-758e-4dcc-a7eb-74fe410cf52d" (UID: "0ff99867-758e-4dcc-a7eb-74fe410cf52d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:04:45 crc kubenswrapper[4758]: I1004 11:04:45.977815 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ff99867-758e-4dcc-a7eb-74fe410cf52d-kube-api-access-d2566" (OuterVolumeSpecName: "kube-api-access-d2566") pod "0ff99867-758e-4dcc-a7eb-74fe410cf52d" (UID: "0ff99867-758e-4dcc-a7eb-74fe410cf52d"). InnerVolumeSpecName "kube-api-access-d2566". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.015359 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ff99867-758e-4dcc-a7eb-74fe410cf52d" (UID: "0ff99867-758e-4dcc-a7eb-74fe410cf52d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.074326 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2566\" (UniqueName: \"kubernetes.io/projected/0ff99867-758e-4dcc-a7eb-74fe410cf52d-kube-api-access-d2566\") on node \"crc\" DevicePath \"\"" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.074363 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.074372 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff99867-758e-4dcc-a7eb-74fe410cf52d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.817383 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24p8j" event={"ID":"0ff99867-758e-4dcc-a7eb-74fe410cf52d","Type":"ContainerDied","Data":"7efb3a1785c037c91a7c7215f9e38545e99b8975b3b4a79427c2a91bb6b5891c"} Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.817422 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24p8j" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.817448 4758 scope.go:117] "RemoveContainer" containerID="71c46a8a6a1a7cbf5c517f918dd5386a874f02b70e2b923c6a0744928ee27e92" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.835893 4758 scope.go:117] "RemoveContainer" containerID="48ebb121aea7cb2bba239a1f5b9ccb954b7d8577fd04f8b0656ab5098b2e27db" Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.851579 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24p8j"] Oct 04 11:04:46 crc kubenswrapper[4758]: I1004 11:04:46.857030 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-24p8j"] Oct 04 11:04:47 crc kubenswrapper[4758]: I1004 11:04:47.331328 4758 scope.go:117] "RemoveContainer" containerID="114123af40c52f01ec336dc7fe1b0c0e1f45880e50690ef81506a1494c381a26" Oct 04 11:04:47 crc kubenswrapper[4758]: I1004 11:04:47.337951 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" path="/var/lib/kubelet/pods/0ff99867-758e-4dcc-a7eb-74fe410cf52d/volumes" Oct 04 11:04:47 crc kubenswrapper[4758]: I1004 11:04:47.826988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" event={"ID":"5b89709e-0a88-4e46-86e6-09e5b1b604da","Type":"ContainerStarted","Data":"0e0e6d601b9c45f6e00aff072be8ffa9b1d2ee02cc518a37942403bc95305d7c"} Oct 04 11:04:47 crc kubenswrapper[4758]: I1004 11:04:47.828118 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:04:47 crc kubenswrapper[4758]: I1004 11:04:47.853050 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" podStartSLOduration=2.904368072 podStartE2EDuration="59.853036738s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.384051113 +0000 UTC m=+807.676702002" lastFinishedPulling="2025-10-04 11:04:47.332719779 +0000 UTC m=+864.625370668" observedRunningTime="2025-10-04 11:04:47.848346232 +0000 UTC m=+865.140997141" watchObservedRunningTime="2025-10-04 11:04:47.853036738 +0000 UTC m=+865.145687637" Oct 04 11:04:52 crc kubenswrapper[4758]: I1004 11:04:52.862746 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" event={"ID":"e355cfb0-5886-4e4a-a1cd-48b6e5b442f6","Type":"ContainerStarted","Data":"924bcaddee06afade9c3f3df8840b2ee6e0092e0898bba1dd888182290fb7a9a"} Oct 04 11:04:52 crc kubenswrapper[4758]: I1004 11:04:52.880260 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-79d8469568-zxtqn" podStartSLOduration=3.212843897 podStartE2EDuration="1m4.880241424s" podCreationTimestamp="2025-10-04 11:03:48 +0000 UTC" firstStartedPulling="2025-10-04 11:03:50.396992855 +0000 UTC m=+807.689643754" lastFinishedPulling="2025-10-04 11:04:52.064390392 +0000 UTC m=+869.357041281" observedRunningTime="2025-10-04 11:04:52.878676932 +0000 UTC m=+870.171327831" watchObservedRunningTime="2025-10-04 11:04:52.880241424 +0000 UTC m=+870.172892313" Oct 04 11:04:55 crc kubenswrapper[4758]: I1004 11:04:55.072468 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d2vpw" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" probeResult="failure" output=< Oct 04 11:04:55 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:04:55 crc kubenswrapper[4758]: > Oct 04 11:04:59 crc kubenswrapper[4758]: I1004 11:04:59.463449 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-647c7ff67btjqdt" Oct 04 11:05:04 crc kubenswrapper[4758]: I1004 11:05:04.071729 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:05:04 crc kubenswrapper[4758]: I1004 11:05:04.116736 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:05:04 crc kubenswrapper[4758]: I1004 11:05:04.500413 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2vpw"] Oct 04 11:05:05 crc kubenswrapper[4758]: I1004 11:05:05.962878 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d2vpw" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" containerID="cri-o://fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2" gracePeriod=2 Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.363915 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.462715 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-utilities\") pod \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.462898 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-catalog-content\") pod \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.463527 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-utilities" (OuterVolumeSpecName: "utilities") pod "0d3122ac-f3b9-49c3-96fa-58109aa83dc6" (UID: "0d3122ac-f3b9-49c3-96fa-58109aa83dc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.464197 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7gzz\" (UniqueName: \"kubernetes.io/projected/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-kube-api-access-v7gzz\") pod \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\" (UID: \"0d3122ac-f3b9-49c3-96fa-58109aa83dc6\") " Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.464568 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.470512 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-kube-api-access-v7gzz" (OuterVolumeSpecName: "kube-api-access-v7gzz") pod "0d3122ac-f3b9-49c3-96fa-58109aa83dc6" (UID: "0d3122ac-f3b9-49c3-96fa-58109aa83dc6"). InnerVolumeSpecName "kube-api-access-v7gzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.559741 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d3122ac-f3b9-49c3-96fa-58109aa83dc6" (UID: "0d3122ac-f3b9-49c3-96fa-58109aa83dc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.566261 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.566302 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7gzz\" (UniqueName: \"kubernetes.io/projected/0d3122ac-f3b9-49c3-96fa-58109aa83dc6-kube-api-access-v7gzz\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.973498 4758 generic.go:334] "Generic (PLEG): container finished" podID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerID="fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2" exitCode=0 Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.973535 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d2vpw" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.973559 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerDied","Data":"fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2"} Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.974925 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d2vpw" event={"ID":"0d3122ac-f3b9-49c3-96fa-58109aa83dc6","Type":"ContainerDied","Data":"9addf9f909b43f5bcfe85919959db029d6f084c5ecc040244c0f1144b2be9d9c"} Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.974950 4758 scope.go:117] "RemoveContainer" containerID="fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2" Oct 04 11:05:06 crc kubenswrapper[4758]: I1004 11:05:06.995357 4758 scope.go:117] "RemoveContainer" containerID="211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.020771 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d2vpw"] Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.023218 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d2vpw"] Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.023376 4758 scope.go:117] "RemoveContainer" containerID="e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.055349 4758 scope.go:117] "RemoveContainer" containerID="fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2" Oct 04 11:05:07 crc kubenswrapper[4758]: E1004 11:05:07.055778 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2\": container with ID starting with fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2 not found: ID does not exist" containerID="fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.055826 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2"} err="failed to get container status \"fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2\": rpc error: code = NotFound desc = could not find container \"fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2\": container with ID starting with fca448840b39f30ebd253e5fe5fd8bfa3ec4d509f0356ba4a79bb5eb1505b6a2 not found: ID does not exist" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.055861 4758 scope.go:117] "RemoveContainer" containerID="211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c" Oct 04 11:05:07 crc kubenswrapper[4758]: E1004 11:05:07.056928 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c\": container with ID starting with 211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c not found: ID does not exist" containerID="211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.056956 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c"} err="failed to get container status \"211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c\": rpc error: code = NotFound desc = could not find container \"211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c\": container with ID starting with 211af5ff72b1bcd5fedc98c8e71097eba3be6ac67b0d72e3d8d319d8c05aaa2c not found: ID does not exist" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.056996 4758 scope.go:117] "RemoveContainer" containerID="e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357" Oct 04 11:05:07 crc kubenswrapper[4758]: E1004 11:05:07.057307 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357\": container with ID starting with e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357 not found: ID does not exist" containerID="e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.057338 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357"} err="failed to get container status \"e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357\": rpc error: code = NotFound desc = could not find container \"e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357\": container with ID starting with e0f37fe42ba80ec52ad17b9569322ee7dd9752b1bd7fdf10ccb00306fe6b3357 not found: ID does not exist" Oct 04 11:05:07 crc kubenswrapper[4758]: I1004 11:05:07.336493 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" path="/var/lib/kubelet/pods/0d3122ac-f3b9-49c3-96fa-58109aa83dc6/volumes" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.610295 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7wp9v"] Oct 04 11:05:15 crc kubenswrapper[4758]: E1004 11:05:15.611208 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="extract-content" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611223 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="extract-content" Oct 04 11:05:15 crc kubenswrapper[4758]: E1004 11:05:15.611242 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="extract-utilities" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611249 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="extract-utilities" Oct 04 11:05:15 crc kubenswrapper[4758]: E1004 11:05:15.611268 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="extract-utilities" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611274 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="extract-utilities" Oct 04 11:05:15 crc kubenswrapper[4758]: E1004 11:05:15.611285 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="extract-content" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611291 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="extract-content" Oct 04 11:05:15 crc kubenswrapper[4758]: E1004 11:05:15.611302 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611307 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" Oct 04 11:05:15 crc kubenswrapper[4758]: E1004 11:05:15.611324 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="registry-server" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611329 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="registry-server" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611476 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3122ac-f3b9-49c3-96fa-58109aa83dc6" containerName="registry-server" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.611489 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ff99867-758e-4dcc-a7eb-74fe410cf52d" containerName="registry-server" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.612158 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.614355 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.614555 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-92vmw" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.614765 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.615095 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.628086 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7wp9v"] Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.688497 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq6qh\" (UniqueName: \"kubernetes.io/projected/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-kube-api-access-kq6qh\") pod \"dnsmasq-dns-675f4bcbfc-7wp9v\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.688567 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-config\") pod \"dnsmasq-dns-675f4bcbfc-7wp9v\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.699918 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zvqfq"] Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.701119 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.705896 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.716055 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zvqfq"] Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.790272 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.790320 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq6qh\" (UniqueName: \"kubernetes.io/projected/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-kube-api-access-kq6qh\") pod \"dnsmasq-dns-675f4bcbfc-7wp9v\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.790377 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-config\") pod \"dnsmasq-dns-675f4bcbfc-7wp9v\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.790405 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-config\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.790431 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bq2v\" (UniqueName: \"kubernetes.io/projected/89129939-07e0-45e8-9057-306d9c2a9ee1-kube-api-access-8bq2v\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.791502 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-config\") pod \"dnsmasq-dns-675f4bcbfc-7wp9v\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.811610 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq6qh\" (UniqueName: \"kubernetes.io/projected/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-kube-api-access-kq6qh\") pod \"dnsmasq-dns-675f4bcbfc-7wp9v\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.891323 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-config\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.891374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bq2v\" (UniqueName: \"kubernetes.io/projected/89129939-07e0-45e8-9057-306d9c2a9ee1-kube-api-access-8bq2v\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.891425 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.892248 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.892255 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-config\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.909770 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bq2v\" (UniqueName: \"kubernetes.io/projected/89129939-07e0-45e8-9057-306d9c2a9ee1-kube-api-access-8bq2v\") pod \"dnsmasq-dns-78dd6ddcc-zvqfq\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:15 crc kubenswrapper[4758]: I1004 11:05:15.927076 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:16 crc kubenswrapper[4758]: I1004 11:05:16.014630 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:16 crc kubenswrapper[4758]: I1004 11:05:16.194286 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7wp9v"] Oct 04 11:05:16 crc kubenswrapper[4758]: I1004 11:05:16.214923 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:05:16 crc kubenswrapper[4758]: I1004 11:05:16.505895 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zvqfq"] Oct 04 11:05:17 crc kubenswrapper[4758]: I1004 11:05:17.089787 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" event={"ID":"89129939-07e0-45e8-9057-306d9c2a9ee1","Type":"ContainerStarted","Data":"08e4c8328c2c2bd00604f16ef1a1782f6d00e83ff4bab7d8aa773ecde66d7f66"} Oct 04 11:05:17 crc kubenswrapper[4758]: I1004 11:05:17.091403 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" event={"ID":"f3296b4c-74ad-40ff-8ccf-f64a688bf42d","Type":"ContainerStarted","Data":"42e302049975f88994f04b7b2f9a2c37e814395c23b166c57f58f7dd40b9de2e"} Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.676327 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7wp9v"] Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.711931 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7bv59"] Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.712991 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.727954 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7bv59"] Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.736852 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.736904 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kttd\" (UniqueName: \"kubernetes.io/projected/9f195fce-1d6b-4384-b733-c58ecd3fd16d-kube-api-access-9kttd\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.737025 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-config\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.838472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-config\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.838525 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.838550 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kttd\" (UniqueName: \"kubernetes.io/projected/9f195fce-1d6b-4384-b733-c58ecd3fd16d-kube-api-access-9kttd\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.839639 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-config\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.840671 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:18 crc kubenswrapper[4758]: I1004 11:05:18.869005 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kttd\" (UniqueName: \"kubernetes.io/projected/9f195fce-1d6b-4384-b733-c58ecd3fd16d-kube-api-access-9kttd\") pod \"dnsmasq-dns-666b6646f7-7bv59\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.002581 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zvqfq"] Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.031030 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hhg4h"] Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.032255 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.035948 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.041512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m9lp\" (UniqueName: \"kubernetes.io/projected/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-kube-api-access-7m9lp\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.041592 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-config\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.045704 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hhg4h"] Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.048850 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.156525 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-config\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.156595 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.156636 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m9lp\" (UniqueName: \"kubernetes.io/projected/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-kube-api-access-7m9lp\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.157500 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.157515 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-config\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.179607 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m9lp\" (UniqueName: \"kubernetes.io/projected/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-kube-api-access-7m9lp\") pod \"dnsmasq-dns-57d769cc4f-hhg4h\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.358480 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.665523 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7bv59"] Oct 04 11:05:19 crc kubenswrapper[4758]: W1004 11:05:19.684572 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f195fce_1d6b_4384_b733_c58ecd3fd16d.slice/crio-455ea644323ecb3e0aeaec086ace4405a9b8012adb85d6893bb78c9d63086058 WatchSource:0}: Error finding container 455ea644323ecb3e0aeaec086ace4405a9b8012adb85d6893bb78c9d63086058: Status 404 returned error can't find the container with id 455ea644323ecb3e0aeaec086ace4405a9b8012adb85d6893bb78c9d63086058 Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.715576 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hhg4h"] Oct 04 11:05:19 crc kubenswrapper[4758]: W1004 11:05:19.724484 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf24fc3d7_4756_4abb_bcd7_2550349ef1ea.slice/crio-14d54f49460a4a1b22763b30cfb4594fd103566774f1fe351d809be204703a1d WatchSource:0}: Error finding container 14d54f49460a4a1b22763b30cfb4594fd103566774f1fe351d809be204703a1d: Status 404 returned error can't find the container with id 14d54f49460a4a1b22763b30cfb4594fd103566774f1fe351d809be204703a1d Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.897872 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.899144 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.901010 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.901432 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.901602 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.903808 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.904276 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.904495 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.904919 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ctr8l" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.918158 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978074 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978155 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978218 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978240 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978266 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978479 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978530 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/acd7d674-14dc-4292-a4b1-5175299225e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978560 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978616 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/acd7d674-14dc-4292-a4b1-5175299225e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978643 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hbxs\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-kube-api-access-8hbxs\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:19 crc kubenswrapper[4758]: I1004 11:05:19.978659 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080460 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080488 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080514 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080535 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/acd7d674-14dc-4292-a4b1-5175299225e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080554 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080577 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/acd7d674-14dc-4292-a4b1-5175299225e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080604 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hbxs\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-kube-api-access-8hbxs\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.080682 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.081620 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.081918 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.082266 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.082852 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.084874 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.082981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-config-data\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.087284 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/acd7d674-14dc-4292-a4b1-5175299225e0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.088161 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.101009 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.113927 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/acd7d674-14dc-4292-a4b1-5175299225e0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.117259 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hbxs\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-kube-api-access-8hbxs\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.122027 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" event={"ID":"9f195fce-1d6b-4384-b733-c58ecd3fd16d","Type":"ContainerStarted","Data":"455ea644323ecb3e0aeaec086ace4405a9b8012adb85d6893bb78c9d63086058"} Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.123653 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" event={"ID":"f24fc3d7-4756-4abb-bcd7-2550349ef1ea","Type":"ContainerStarted","Data":"14d54f49460a4a1b22763b30cfb4594fd103566774f1fe351d809be204703a1d"} Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.129984 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.189528 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.191088 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.193544 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.194427 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.194665 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.194969 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gvv74" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.195864 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.196534 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.201530 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.205638 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.227224 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.287508 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289363 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57d6ac32-db38-4443-9c0c-13c66db2451b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289520 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289553 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m7kn\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-kube-api-access-5m7kn\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289592 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289658 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289836 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.289933 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.290007 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57d6ac32-db38-4443-9c0c-13c66db2451b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.290027 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.290054 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57d6ac32-db38-4443-9c0c-13c66db2451b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391410 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391501 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391695 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391742 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57d6ac32-db38-4443-9c0c-13c66db2451b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391766 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391788 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m7kn\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-kube-api-access-5m7kn\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391869 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391896 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.391985 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.393449 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.393648 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.393812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.401902 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.402088 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.402268 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.431176 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.435271 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57d6ac32-db38-4443-9c0c-13c66db2451b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.435675 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57d6ac32-db38-4443-9c0c-13c66db2451b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.435738 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.436043 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.439452 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m7kn\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-kube-api-access-5m7kn\") pod \"rabbitmq-cell1-server-0\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.507365 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.720134 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:05:20 crc kubenswrapper[4758]: W1004 11:05:20.783317 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacd7d674_14dc_4292_a4b1_5175299225e0.slice/crio-354b9d7f44922144c326175be25e0116e0de5eee93bfe37729b5ad5ee9f3aba6 WatchSource:0}: Error finding container 354b9d7f44922144c326175be25e0116e0de5eee93bfe37729b5ad5ee9f3aba6: Status 404 returned error can't find the container with id 354b9d7f44922144c326175be25e0116e0de5eee93bfe37729b5ad5ee9f3aba6 Oct 04 11:05:20 crc kubenswrapper[4758]: I1004 11:05:20.956628 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:05:20 crc kubenswrapper[4758]: W1004 11:05:20.974825 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57d6ac32_db38_4443_9c0c_13c66db2451b.slice/crio-ecb705bc249ee000fc5a802b33bccf037c245032292d7fac85d80c388236539e WatchSource:0}: Error finding container ecb705bc249ee000fc5a802b33bccf037c245032292d7fac85d80c388236539e: Status 404 returned error can't find the container with id ecb705bc249ee000fc5a802b33bccf037c245032292d7fac85d80c388236539e Oct 04 11:05:21 crc kubenswrapper[4758]: I1004 11:05:21.154670 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"acd7d674-14dc-4292-a4b1-5175299225e0","Type":"ContainerStarted","Data":"354b9d7f44922144c326175be25e0116e0de5eee93bfe37729b5ad5ee9f3aba6"} Oct 04 11:05:21 crc kubenswrapper[4758]: I1004 11:05:21.200026 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"57d6ac32-db38-4443-9c0c-13c66db2451b","Type":"ContainerStarted","Data":"ecb705bc249ee000fc5a802b33bccf037c245032292d7fac85d80c388236539e"} Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.328619 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.330218 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.335630 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.336543 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.339132 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-vcpwh" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.339278 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.339404 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.342589 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.361882 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445616 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445720 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445780 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-secrets\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445805 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445878 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445911 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445931 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445969 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.445989 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m5ll\" (UniqueName: \"kubernetes.io/projected/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-kube-api-access-6m5ll\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.548793 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.548872 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.548930 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-secrets\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.548978 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.549019 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.549042 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.549056 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.549090 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.549538 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.549947 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.550502 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m5ll\" (UniqueName: \"kubernetes.io/projected/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-kube-api-access-6m5ll\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.551323 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.551370 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.551795 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.564838 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-secrets\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.565813 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.567464 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.571411 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m5ll\" (UniqueName: \"kubernetes.io/projected/ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a-kube-api-access-6m5ll\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.577122 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a\") " pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.657458 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.659474 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.662251 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-d9gxg" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.666044 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.666385 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.667314 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.669962 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.680275 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.755604 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.755693 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.755751 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.755793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.755870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.755928 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.756034 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.756091 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fgw\" (UniqueName: \"kubernetes.io/projected/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-kube-api-access-d8fgw\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.756223 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.857928 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.857990 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fgw\" (UniqueName: \"kubernetes.io/projected/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-kube-api-access-d8fgw\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858025 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858640 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858681 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858703 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858731 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858774 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.858802 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.859427 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.860739 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.861377 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.861536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.862891 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.870854 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.871211 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.880090 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.881670 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fgw\" (UniqueName: \"kubernetes.io/projected/3d86d897-f3e3-4eac-94f4-6fe59fc58d20-kube-api-access-d8fgw\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:22 crc kubenswrapper[4758]: I1004 11:05:22.937551 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3d86d897-f3e3-4eac-94f4-6fe59fc58d20\") " pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.002947 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.314204 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.322350 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.324733 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-dxcth" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.324890 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.334735 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.354710 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.371258 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9488150-f848-482f-8397-5ec145f84af3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.371305 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxj9g\" (UniqueName: \"kubernetes.io/projected/c9488150-f848-482f-8397-5ec145f84af3-kube-api-access-cxj9g\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.371348 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9488150-f848-482f-8397-5ec145f84af3-kolla-config\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.371364 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9488150-f848-482f-8397-5ec145f84af3-config-data\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.371378 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9488150-f848-482f-8397-5ec145f84af3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.472194 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9488150-f848-482f-8397-5ec145f84af3-kolla-config\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.482214 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9488150-f848-482f-8397-5ec145f84af3-config-data\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.482279 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9488150-f848-482f-8397-5ec145f84af3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.482560 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9488150-f848-482f-8397-5ec145f84af3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.482610 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxj9g\" (UniqueName: \"kubernetes.io/projected/c9488150-f848-482f-8397-5ec145f84af3-kube-api-access-cxj9g\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.477470 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c9488150-f848-482f-8397-5ec145f84af3-kolla-config\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.483748 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c9488150-f848-482f-8397-5ec145f84af3-config-data\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.491765 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9488150-f848-482f-8397-5ec145f84af3-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.499844 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c9488150-f848-482f-8397-5ec145f84af3-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.505508 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxj9g\" (UniqueName: \"kubernetes.io/projected/c9488150-f848-482f-8397-5ec145f84af3-kube-api-access-cxj9g\") pod \"memcached-0\" (UID: \"c9488150-f848-482f-8397-5ec145f84af3\") " pod="openstack/memcached-0" Oct 04 11:05:23 crc kubenswrapper[4758]: I1004 11:05:23.668899 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.210064 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.210959 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.213312 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-zcjrx" Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.276778 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.292960 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6vbp\" (UniqueName: \"kubernetes.io/projected/f3b5b5a4-cf2e-489c-acd3-61811589fa03-kube-api-access-f6vbp\") pod \"kube-state-metrics-0\" (UID: \"f3b5b5a4-cf2e-489c-acd3-61811589fa03\") " pod="openstack/kube-state-metrics-0" Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.394212 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6vbp\" (UniqueName: \"kubernetes.io/projected/f3b5b5a4-cf2e-489c-acd3-61811589fa03-kube-api-access-f6vbp\") pod \"kube-state-metrics-0\" (UID: \"f3b5b5a4-cf2e-489c-acd3-61811589fa03\") " pod="openstack/kube-state-metrics-0" Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.409997 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6vbp\" (UniqueName: \"kubernetes.io/projected/f3b5b5a4-cf2e-489c-acd3-61811589fa03-kube-api-access-f6vbp\") pod \"kube-state-metrics-0\" (UID: \"f3b5b5a4-cf2e-489c-acd3-61811589fa03\") " pod="openstack/kube-state-metrics-0" Oct 04 11:05:24 crc kubenswrapper[4758]: I1004 11:05:24.529127 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.630382 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gpsfr"] Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.633790 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.642659 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-mlrts" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.642830 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.643092 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.655918 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gpsfr"] Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.704853 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7fwv4"] Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.706504 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.719361 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7fwv4"] Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.756722 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-lib\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.756856 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-run\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.756891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39bf6bd9-f043-4485-8980-846341f31b4e-scripts\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.756917 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vqkq\" (UniqueName: \"kubernetes.io/projected/e4d252bc-617d-4014-b340-ce7feb1c7062-kube-api-access-2vqkq\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757002 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bf6bd9-f043-4485-8980-846341f31b4e-combined-ca-bundle\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757212 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-log\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-run-ovn\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757256 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-log-ovn\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757295 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bf6bd9-f043-4485-8980-846341f31b4e-ovn-controller-tls-certs\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757350 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4d252bc-617d-4014-b340-ce7feb1c7062-scripts\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757373 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-run\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757402 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-etc-ovs\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.757422 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjchj\" (UniqueName: \"kubernetes.io/projected/39bf6bd9-f043-4485-8980-846341f31b4e-kube-api-access-jjchj\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859022 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-log\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859061 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-run-ovn\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859079 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-log-ovn\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859118 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bf6bd9-f043-4485-8980-846341f31b4e-ovn-controller-tls-certs\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859155 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4d252bc-617d-4014-b340-ce7feb1c7062-scripts\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859295 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-run\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-etc-ovs\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859347 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjchj\" (UniqueName: \"kubernetes.io/projected/39bf6bd9-f043-4485-8980-846341f31b4e-kube-api-access-jjchj\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859372 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-lib\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859392 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-run\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39bf6bd9-f043-4485-8980-846341f31b4e-scripts\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859427 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vqkq\" (UniqueName: \"kubernetes.io/projected/e4d252bc-617d-4014-b340-ce7feb1c7062-kube-api-access-2vqkq\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859448 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bf6bd9-f043-4485-8980-846341f31b4e-combined-ca-bundle\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859512 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-log\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859633 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-log-ovn\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859690 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-run-ovn\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859856 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-etc-ovs\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.859946 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/39bf6bd9-f043-4485-8980-846341f31b4e-var-run\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.860019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-run\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.860642 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e4d252bc-617d-4014-b340-ce7feb1c7062-var-lib\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.861754 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39bf6bd9-f043-4485-8980-846341f31b4e-scripts\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.863229 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e4d252bc-617d-4014-b340-ce7feb1c7062-scripts\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.865272 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/39bf6bd9-f043-4485-8980-846341f31b4e-ovn-controller-tls-certs\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.866771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39bf6bd9-f043-4485-8980-846341f31b4e-combined-ca-bundle\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.883097 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjchj\" (UniqueName: \"kubernetes.io/projected/39bf6bd9-f043-4485-8980-846341f31b4e-kube-api-access-jjchj\") pod \"ovn-controller-gpsfr\" (UID: \"39bf6bd9-f043-4485-8980-846341f31b4e\") " pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.884753 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vqkq\" (UniqueName: \"kubernetes.io/projected/e4d252bc-617d-4014-b340-ce7feb1c7062-kube-api-access-2vqkq\") pod \"ovn-controller-ovs-7fwv4\" (UID: \"e4d252bc-617d-4014-b340-ce7feb1c7062\") " pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:27 crc kubenswrapper[4758]: I1004 11:05:27.954425 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.021617 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.528556 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.529724 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.532267 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-fbqlh" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.532377 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.532434 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.533709 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.575628 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmtmj\" (UniqueName: \"kubernetes.io/projected/1c45f9ea-ff92-44d1-801f-2f0faf16597d-kube-api-access-pmtmj\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.575677 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c45f9ea-ff92-44d1-801f-2f0faf16597d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.575698 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c45f9ea-ff92-44d1-801f-2f0faf16597d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.575716 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c45f9ea-ff92-44d1-801f-2f0faf16597d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.575753 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.677551 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmtmj\" (UniqueName: \"kubernetes.io/projected/1c45f9ea-ff92-44d1-801f-2f0faf16597d-kube-api-access-pmtmj\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.677609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c45f9ea-ff92-44d1-801f-2f0faf16597d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.677638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c45f9ea-ff92-44d1-801f-2f0faf16597d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.677660 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c45f9ea-ff92-44d1-801f-2f0faf16597d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.677694 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.678068 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.680626 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c45f9ea-ff92-44d1-801f-2f0faf16597d-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.695763 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c45f9ea-ff92-44d1-801f-2f0faf16597d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.696468 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmtmj\" (UniqueName: \"kubernetes.io/projected/1c45f9ea-ff92-44d1-801f-2f0faf16597d-kube-api-access-pmtmj\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.709793 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c45f9ea-ff92-44d1-801f-2f0faf16597d-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.722567 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1c45f9ea-ff92-44d1-801f-2f0faf16597d\") " pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:28 crc kubenswrapper[4758]: I1004 11:05:28.852746 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.529646 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.532924 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.540248 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.542660 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.544185 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-6sqvn" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.573240 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.626263 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjzgm\" (UniqueName: \"kubernetes.io/projected/8a3affc8-eb3e-4ab1-ab10-07b690442cff-kube-api-access-fjzgm\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.626301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.626428 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3affc8-eb3e-4ab1-ab10-07b690442cff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.626465 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a3affc8-eb3e-4ab1-ab10-07b690442cff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.626496 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3affc8-eb3e-4ab1-ab10-07b690442cff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.727588 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjzgm\" (UniqueName: \"kubernetes.io/projected/8a3affc8-eb3e-4ab1-ab10-07b690442cff-kube-api-access-fjzgm\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.727636 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.727681 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3affc8-eb3e-4ab1-ab10-07b690442cff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.727766 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a3affc8-eb3e-4ab1-ab10-07b690442cff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.727801 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3affc8-eb3e-4ab1-ab10-07b690442cff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.729066 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.731315 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a3affc8-eb3e-4ab1-ab10-07b690442cff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.741418 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a3affc8-eb3e-4ab1-ab10-07b690442cff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.742324 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a3affc8-eb3e-4ab1-ab10-07b690442cff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.749671 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjzgm\" (UniqueName: \"kubernetes.io/projected/8a3affc8-eb3e-4ab1-ab10-07b690442cff-kube-api-access-fjzgm\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.753585 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8a3affc8-eb3e-4ab1-ab10-07b690442cff\") " pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:31 crc kubenswrapper[4758]: I1004 11:05:31.859986 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.948393 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.949300 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8bq2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-zvqfq_openstack(89129939-07e0-45e8-9057-306d9c2a9ee1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.950470 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" podUID="89129939-07e0-45e8-9057-306d9c2a9ee1" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.991087 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.991337 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kq6qh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-7wp9v_openstack(f3296b4c-74ad-40ff-8ccf-f64a688bf42d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.993430 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" podUID="f3296b4c-74ad-40ff-8ccf-f64a688bf42d" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.999749 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 11:05:41 crc kubenswrapper[4758]: E1004 11:05:41.999863 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7m9lp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-hhg4h_openstack(f24fc3d7-4756-4abb-bcd7-2550349ef1ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:05:42 crc kubenswrapper[4758]: E1004 11:05:42.001107 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" podUID="f24fc3d7-4756-4abb-bcd7-2550349ef1ea" Oct 04 11:05:42 crc kubenswrapper[4758]: E1004 11:05:42.015033 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 11:05:42 crc kubenswrapper[4758]: E1004 11:05:42.015358 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9kttd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-7bv59_openstack(9f195fce-1d6b-4384-b733-c58ecd3fd16d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:05:42 crc kubenswrapper[4758]: E1004 11:05:42.016429 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" podUID="9f195fce-1d6b-4384-b733-c58ecd3fd16d" Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.293734 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:05:42 crc kubenswrapper[4758]: W1004 11:05:42.295945 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3b5b5a4_cf2e_489c_acd3_61811589fa03.slice/crio-6b3957ceff5117d07881b1bf016c8348b10faa0a76b3759074dc947d88b924a0 WatchSource:0}: Error finding container 6b3957ceff5117d07881b1bf016c8348b10faa0a76b3759074dc947d88b924a0: Status 404 returned error can't find the container with id 6b3957ceff5117d07881b1bf016c8348b10faa0a76b3759074dc947d88b924a0 Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.398015 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"57d6ac32-db38-4443-9c0c-13c66db2451b","Type":"ContainerStarted","Data":"a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17"} Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.399731 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3b5b5a4-cf2e-489c-acd3-61811589fa03","Type":"ContainerStarted","Data":"6b3957ceff5117d07881b1bf016c8348b10faa0a76b3759074dc947d88b924a0"} Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.400907 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"acd7d674-14dc-4292-a4b1-5175299225e0","Type":"ContainerStarted","Data":"fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4"} Oct 04 11:05:42 crc kubenswrapper[4758]: E1004 11:05:42.405559 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" podUID="9f195fce-1d6b-4384-b733-c58ecd3fd16d" Oct 04 11:05:42 crc kubenswrapper[4758]: E1004 11:05:42.405669 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" podUID="f24fc3d7-4756-4abb-bcd7-2550349ef1ea" Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.720964 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.858081 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.870142 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.886710 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.908963 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gpsfr"] Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.909640 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bq2v\" (UniqueName: \"kubernetes.io/projected/89129939-07e0-45e8-9057-306d9c2a9ee1-kube-api-access-8bq2v\") pod \"89129939-07e0-45e8-9057-306d9c2a9ee1\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.909707 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-dns-svc\") pod \"89129939-07e0-45e8-9057-306d9c2a9ee1\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.909761 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-config\") pod \"89129939-07e0-45e8-9057-306d9c2a9ee1\" (UID: \"89129939-07e0-45e8-9057-306d9c2a9ee1\") " Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.910368 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89129939-07e0-45e8-9057-306d9c2a9ee1" (UID: "89129939-07e0-45e8-9057-306d9c2a9ee1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.922391 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-config" (OuterVolumeSpecName: "config") pod "89129939-07e0-45e8-9057-306d9c2a9ee1" (UID: "89129939-07e0-45e8-9057-306d9c2a9ee1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:42 crc kubenswrapper[4758]: I1004 11:05:42.927349 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89129939-07e0-45e8-9057-306d9c2a9ee1-kube-api-access-8bq2v" (OuterVolumeSpecName: "kube-api-access-8bq2v") pod "89129939-07e0-45e8-9057-306d9c2a9ee1" (UID: "89129939-07e0-45e8-9057-306d9c2a9ee1"). InnerVolumeSpecName "kube-api-access-8bq2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.003355 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.005991 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7fwv4"] Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.011732 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.011763 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89129939-07e0-45e8-9057-306d9c2a9ee1-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.011776 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bq2v\" (UniqueName: \"kubernetes.io/projected/89129939-07e0-45e8-9057-306d9c2a9ee1-kube-api-access-8bq2v\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.112660 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq6qh\" (UniqueName: \"kubernetes.io/projected/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-kube-api-access-kq6qh\") pod \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.112881 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-config\") pod \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\" (UID: \"f3296b4c-74ad-40ff-8ccf-f64a688bf42d\") " Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.113376 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-config" (OuterVolumeSpecName: "config") pod "f3296b4c-74ad-40ff-8ccf-f64a688bf42d" (UID: "f3296b4c-74ad-40ff-8ccf-f64a688bf42d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.116336 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-kube-api-access-kq6qh" (OuterVolumeSpecName: "kube-api-access-kq6qh") pod "f3296b4c-74ad-40ff-8ccf-f64a688bf42d" (UID: "f3296b4c-74ad-40ff-8ccf-f64a688bf42d"). InnerVolumeSpecName "kube-api-access-kq6qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.214393 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.214428 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq6qh\" (UniqueName: \"kubernetes.io/projected/f3296b4c-74ad-40ff-8ccf-f64a688bf42d-kube-api-access-kq6qh\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.406987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" event={"ID":"89129939-07e0-45e8-9057-306d9c2a9ee1","Type":"ContainerDied","Data":"08e4c8328c2c2bd00604f16ef1a1782f6d00e83ff4bab7d8aa773ecde66d7f66"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.407223 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-zvqfq" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.408301 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c9488150-f848-482f-8397-5ec145f84af3","Type":"ContainerStarted","Data":"5c1f71a3698392b7e5d4bfed74e7c474ac6fbbc3a7404f35d0d883d7956b2088"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.409011 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gpsfr" event={"ID":"39bf6bd9-f043-4485-8980-846341f31b4e","Type":"ContainerStarted","Data":"3bd0c09269c14a835c4732f4ea832471bc05c8772345dafe5e8e8f63e382d62a"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.414690 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fwv4" event={"ID":"e4d252bc-617d-4014-b340-ce7feb1c7062","Type":"ContainerStarted","Data":"7736a73419e9678f63aed0a82e6cbff22681e5719cd54c519d64d50e01da0f50"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.418361 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a","Type":"ContainerStarted","Data":"b50c37ce07ad6e6848afbde83f8a777831ac99e31ae3e0bd3f49f30a8007b67d"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.419734 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" event={"ID":"f3296b4c-74ad-40ff-8ccf-f64a688bf42d","Type":"ContainerDied","Data":"42e302049975f88994f04b7b2f9a2c37e814395c23b166c57f58f7dd40b9de2e"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.419804 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7wp9v" Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.420974 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3d86d897-f3e3-4eac-94f4-6fe59fc58d20","Type":"ContainerStarted","Data":"6ef7f972084785052478486f8d4d69a96960a58c1e44389798a73e3a648d4c68"} Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.503256 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zvqfq"] Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.525903 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-zvqfq"] Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.540918 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7wp9v"] Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.545276 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7wp9v"] Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.579667 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 11:05:43 crc kubenswrapper[4758]: I1004 11:05:43.751950 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 11:05:44 crc kubenswrapper[4758]: W1004 11:05:44.924852 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c45f9ea_ff92_44d1_801f_2f0faf16597d.slice/crio-e5a12c537c89eee9d85eb602c7edaa62d32b74fad9b830e0d2dffee3972f0085 WatchSource:0}: Error finding container e5a12c537c89eee9d85eb602c7edaa62d32b74fad9b830e0d2dffee3972f0085: Status 404 returned error can't find the container with id e5a12c537c89eee9d85eb602c7edaa62d32b74fad9b830e0d2dffee3972f0085 Oct 04 11:05:45 crc kubenswrapper[4758]: W1004 11:05:45.014376 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a3affc8_eb3e_4ab1_ab10_07b690442cff.slice/crio-d44491cc048317369ad59801314edb2493dd97d6168ba3cb18e52ef56df141d5 WatchSource:0}: Error finding container d44491cc048317369ad59801314edb2493dd97d6168ba3cb18e52ef56df141d5: Status 404 returned error can't find the container with id d44491cc048317369ad59801314edb2493dd97d6168ba3cb18e52ef56df141d5 Oct 04 11:05:45 crc kubenswrapper[4758]: I1004 11:05:45.335654 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89129939-07e0-45e8-9057-306d9c2a9ee1" path="/var/lib/kubelet/pods/89129939-07e0-45e8-9057-306d9c2a9ee1/volumes" Oct 04 11:05:45 crc kubenswrapper[4758]: I1004 11:05:45.336250 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3296b4c-74ad-40ff-8ccf-f64a688bf42d" path="/var/lib/kubelet/pods/f3296b4c-74ad-40ff-8ccf-f64a688bf42d/volumes" Oct 04 11:05:45 crc kubenswrapper[4758]: I1004 11:05:45.436604 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8a3affc8-eb3e-4ab1-ab10-07b690442cff","Type":"ContainerStarted","Data":"d44491cc048317369ad59801314edb2493dd97d6168ba3cb18e52ef56df141d5"} Oct 04 11:05:45 crc kubenswrapper[4758]: I1004 11:05:45.439160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1c45f9ea-ff92-44d1-801f-2f0faf16597d","Type":"ContainerStarted","Data":"e5a12c537c89eee9d85eb602c7edaa62d32b74fad9b830e0d2dffee3972f0085"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.500267 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3d86d897-f3e3-4eac-94f4-6fe59fc58d20","Type":"ContainerStarted","Data":"91d71dd0010d8acaf73dfb41d65ce5834d50d5b0281498d74073e25683516650"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.501772 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3b5b5a4-cf2e-489c-acd3-61811589fa03","Type":"ContainerStarted","Data":"c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.501872 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.503638 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1c45f9ea-ff92-44d1-801f-2f0faf16597d","Type":"ContainerStarted","Data":"0d747f1f4171f08d315dc0f3b19931dec72e06f507bd643e456001003b4e7ce7"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.504807 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gpsfr" event={"ID":"39bf6bd9-f043-4485-8980-846341f31b4e","Type":"ContainerStarted","Data":"0cbd74397f379c51338af323dee7de6b8b05808aaedda803504c3de708e3b037"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.504987 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gpsfr" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.506011 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c9488150-f848-482f-8397-5ec145f84af3","Type":"ContainerStarted","Data":"c3da6b2f3d1983447ead17bab43a938ba1b45b93397966b92a7b7f01532902cb"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.506147 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.507592 4758 generic.go:334] "Generic (PLEG): container finished" podID="e4d252bc-617d-4014-b340-ce7feb1c7062" containerID="8a162b650a49a18c30b3e1917c5fb66c928daf91ad6f15a28a305b00ccb38e25" exitCode=0 Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.507639 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fwv4" event={"ID":"e4d252bc-617d-4014-b340-ce7feb1c7062","Type":"ContainerDied","Data":"8a162b650a49a18c30b3e1917c5fb66c928daf91ad6f15a28a305b00ccb38e25"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.509154 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a","Type":"ContainerStarted","Data":"2cc57038fbde173648eec8e83c91caffd6cdbdbe169ca2e66dd27e84dbd98175"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.511330 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8a3affc8-eb3e-4ab1-ab10-07b690442cff","Type":"ContainerStarted","Data":"a2a565aac23ea56b575b0e593edb446f8d0e237a31afafee0054c5bf2a677a4b"} Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.542180 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gpsfr" podStartSLOduration=16.878646178 podStartE2EDuration="24.542160093s" podCreationTimestamp="2025-10-04 11:05:27 +0000 UTC" firstStartedPulling="2025-10-04 11:05:42.944935141 +0000 UTC m=+920.237586020" lastFinishedPulling="2025-10-04 11:05:50.608449046 +0000 UTC m=+927.901099935" observedRunningTime="2025-10-04 11:05:51.534847406 +0000 UTC m=+928.827498325" watchObservedRunningTime="2025-10-04 11:05:51.542160093 +0000 UTC m=+928.834810982" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.567807 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.882637525 podStartE2EDuration="28.567785583s" podCreationTimestamp="2025-10-04 11:05:23 +0000 UTC" firstStartedPulling="2025-10-04 11:05:42.87551 +0000 UTC m=+920.168160889" lastFinishedPulling="2025-10-04 11:05:50.560658058 +0000 UTC m=+927.853308947" observedRunningTime="2025-10-04 11:05:51.557057634 +0000 UTC m=+928.849708523" watchObservedRunningTime="2025-10-04 11:05:51.567785583 +0000 UTC m=+928.860436472" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.618438 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=16.028044058 podStartE2EDuration="21.618419048s" podCreationTimestamp="2025-10-04 11:05:30 +0000 UTC" firstStartedPulling="2025-10-04 11:05:45.015864307 +0000 UTC m=+922.308515196" lastFinishedPulling="2025-10-04 11:05:50.606239297 +0000 UTC m=+927.898890186" observedRunningTime="2025-10-04 11:05:51.60996233 +0000 UTC m=+928.902613219" watchObservedRunningTime="2025-10-04 11:05:51.618419048 +0000 UTC m=+928.911069937" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.638149 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.378297937 podStartE2EDuration="27.638127319s" podCreationTimestamp="2025-10-04 11:05:24 +0000 UTC" firstStartedPulling="2025-10-04 11:05:42.301039252 +0000 UTC m=+919.593690141" lastFinishedPulling="2025-10-04 11:05:50.560868634 +0000 UTC m=+927.853519523" observedRunningTime="2025-10-04 11:05:51.631199182 +0000 UTC m=+928.923850071" watchObservedRunningTime="2025-10-04 11:05:51.638127319 +0000 UTC m=+928.930778198" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.668132 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.994199017 podStartE2EDuration="24.668088476s" podCreationTimestamp="2025-10-04 11:05:27 +0000 UTC" firstStartedPulling="2025-10-04 11:05:44.927550668 +0000 UTC m=+922.220201557" lastFinishedPulling="2025-10-04 11:05:50.601440117 +0000 UTC m=+927.894091016" observedRunningTime="2025-10-04 11:05:51.667782558 +0000 UTC m=+928.960433447" watchObservedRunningTime="2025-10-04 11:05:51.668088476 +0000 UTC m=+928.960739365" Oct 04 11:05:51 crc kubenswrapper[4758]: I1004 11:05:51.861058 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:52 crc kubenswrapper[4758]: I1004 11:05:52.522808 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fwv4" event={"ID":"e4d252bc-617d-4014-b340-ce7feb1c7062","Type":"ContainerStarted","Data":"e18ca1d6bc4762744e5f6f9a3140d8c00fc6f1c89ec2582a069856560e17b418"} Oct 04 11:05:52 crc kubenswrapper[4758]: I1004 11:05:52.523153 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fwv4" event={"ID":"e4d252bc-617d-4014-b340-ce7feb1c7062","Type":"ContainerStarted","Data":"07806be53678b0fbd608e27661fc9c6fd818153132e86aacebd1ae5aaa2f372f"} Oct 04 11:05:52 crc kubenswrapper[4758]: I1004 11:05:52.543727 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7fwv4" podStartSLOduration=17.99930456 podStartE2EDuration="25.543711717s" podCreationTimestamp="2025-10-04 11:05:27 +0000 UTC" firstStartedPulling="2025-10-04 11:05:43.022837889 +0000 UTC m=+920.315488778" lastFinishedPulling="2025-10-04 11:05:50.567245046 +0000 UTC m=+927.859895935" observedRunningTime="2025-10-04 11:05:52.542517685 +0000 UTC m=+929.835168584" watchObservedRunningTime="2025-10-04 11:05:52.543711717 +0000 UTC m=+929.836362606" Oct 04 11:05:52 crc kubenswrapper[4758]: I1004 11:05:52.853828 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:52 crc kubenswrapper[4758]: I1004 11:05:52.861214 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:53 crc kubenswrapper[4758]: I1004 11:05:53.023472 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:53 crc kubenswrapper[4758]: I1004 11:05:53.023537 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:05:53 crc kubenswrapper[4758]: I1004 11:05:53.852836 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:54 crc kubenswrapper[4758]: I1004 11:05:54.580231 4758 generic.go:334] "Generic (PLEG): container finished" podID="ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a" containerID="2cc57038fbde173648eec8e83c91caffd6cdbdbe169ca2e66dd27e84dbd98175" exitCode=0 Oct 04 11:05:54 crc kubenswrapper[4758]: I1004 11:05:54.580294 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a","Type":"ContainerDied","Data":"2cc57038fbde173648eec8e83c91caffd6cdbdbe169ca2e66dd27e84dbd98175"} Oct 04 11:05:54 crc kubenswrapper[4758]: I1004 11:05:54.585655 4758 generic.go:334] "Generic (PLEG): container finished" podID="3d86d897-f3e3-4eac-94f4-6fe59fc58d20" containerID="91d71dd0010d8acaf73dfb41d65ce5834d50d5b0281498d74073e25683516650" exitCode=0 Oct 04 11:05:54 crc kubenswrapper[4758]: I1004 11:05:54.586986 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3d86d897-f3e3-4eac-94f4-6fe59fc58d20","Type":"ContainerDied","Data":"91d71dd0010d8acaf73dfb41d65ce5834d50d5b0281498d74073e25683516650"} Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.597520 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a","Type":"ContainerStarted","Data":"11e87321dc9ce26cabfe9d02c6e36e311b4e694e3018a496dfc88a5e25e03c22"} Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.600058 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3d86d897-f3e3-4eac-94f4-6fe59fc58d20","Type":"ContainerStarted","Data":"905283eae62536ef1f9f49f3c5acc1bd41eaa288c2cd75518538668ef3193397"} Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.635471 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=26.944592285 podStartE2EDuration="34.635449497s" podCreationTimestamp="2025-10-04 11:05:21 +0000 UTC" firstStartedPulling="2025-10-04 11:05:42.91632511 +0000 UTC m=+920.208975999" lastFinishedPulling="2025-10-04 11:05:50.607182322 +0000 UTC m=+927.899833211" observedRunningTime="2025-10-04 11:05:55.63071493 +0000 UTC m=+932.923365859" watchObservedRunningTime="2025-10-04 11:05:55.635449497 +0000 UTC m=+932.928100416" Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.657960 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.955951371 podStartE2EDuration="34.657938693s" podCreationTimestamp="2025-10-04 11:05:21 +0000 UTC" firstStartedPulling="2025-10-04 11:05:42.901457649 +0000 UTC m=+920.194108538" lastFinishedPulling="2025-10-04 11:05:50.603444971 +0000 UTC m=+927.896095860" observedRunningTime="2025-10-04 11:05:55.653025581 +0000 UTC m=+932.945676510" watchObservedRunningTime="2025-10-04 11:05:55.657938693 +0000 UTC m=+932.950589592" Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.926859 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.933035 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:55 crc kubenswrapper[4758]: I1004 11:05:55.993990 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.010815 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.233562 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hhg4h"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.297110 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6wnl7"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.298306 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.300914 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.325443 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6wnl7"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.417806 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-9cb54f9cb-xrmhr"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.418830 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.423469 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.423672 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-5cbmc" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.423802 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.428771 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.429122 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-config\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.429193 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.429257 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xbl2\" (UniqueName: \"kubernetes.io/projected/8001680b-56f4-47a4-94e2-891b1a56c2af-kube-api-access-2xbl2\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.476205 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7bv59"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.476275 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-9cb54f9cb-xrmhr"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.512594 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7g8pw"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.515275 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.517243 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532147 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532191 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-config\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532234 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xbl2\" (UniqueName: \"kubernetes.io/projected/8001680b-56f4-47a4-94e2-891b1a56c2af-kube-api-access-2xbl2\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532264 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed03b984-16f3-46dd-a684-f381b6fbe89b-scripts\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532366 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l28q\" (UniqueName: \"kubernetes.io/projected/ed03b984-16f3-46dd-a684-f381b6fbe89b-kube-api-access-4l28q\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532394 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed03b984-16f3-46dd-a684-f381b6fbe89b-combined-ca-bundle\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed03b984-16f3-46dd-a684-f381b6fbe89b-ovn-northd-tls-certs\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.532485 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.533381 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.533764 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.534013 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-config\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.538315 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7g8pw"] Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.553237 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xbl2\" (UniqueName: \"kubernetes.io/projected/8001680b-56f4-47a4-94e2-891b1a56c2af-kube-api-access-2xbl2\") pod \"dnsmasq-dns-7fd796d7df-6wnl7\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.613778 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" event={"ID":"f24fc3d7-4756-4abb-bcd7-2550349ef1ea","Type":"ContainerDied","Data":"14d54f49460a4a1b22763b30cfb4594fd103566774f1fe351d809be204703a1d"} Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.613815 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14d54f49460a4a1b22763b30cfb4594fd103566774f1fe351d809be204703a1d" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.615806 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.623616 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.641834 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.641892 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642230 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642275 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed03b984-16f3-46dd-a684-f381b6fbe89b-scripts\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642305 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l28q\" (UniqueName: \"kubernetes.io/projected/ed03b984-16f3-46dd-a684-f381b6fbe89b-kube-api-access-4l28q\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642330 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed03b984-16f3-46dd-a684-f381b6fbe89b-combined-ca-bundle\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642379 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642406 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj525\" (UniqueName: \"kubernetes.io/projected/6581e45a-6f33-45c5-a98f-c1fe77f6f703-kube-api-access-bj525\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.642427 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed03b984-16f3-46dd-a684-f381b6fbe89b-ovn-northd-tls-certs\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.643904 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed03b984-16f3-46dd-a684-f381b6fbe89b-scripts\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.656825 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed03b984-16f3-46dd-a684-f381b6fbe89b-ovn-northd-tls-certs\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.658811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed03b984-16f3-46dd-a684-f381b6fbe89b-combined-ca-bundle\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.659740 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l28q\" (UniqueName: \"kubernetes.io/projected/ed03b984-16f3-46dd-a684-f381b6fbe89b-kube-api-access-4l28q\") pod \"ovn-northd-9cb54f9cb-xrmhr\" (UID: \"ed03b984-16f3-46dd-a684-f381b6fbe89b\") " pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.744849 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-config\") pod \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745290 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-dns-svc\") pod \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745400 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m9lp\" (UniqueName: \"kubernetes.io/projected/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-kube-api-access-7m9lp\") pod \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\" (UID: \"f24fc3d7-4756-4abb-bcd7-2550349ef1ea\") " Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745510 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-config" (OuterVolumeSpecName: "config") pod "f24fc3d7-4756-4abb-bcd7-2550349ef1ea" (UID: "f24fc3d7-4756-4abb-bcd7-2550349ef1ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745695 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745742 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f24fc3d7-4756-4abb-bcd7-2550349ef1ea" (UID: "f24fc3d7-4756-4abb-bcd7-2550349ef1ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745760 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj525\" (UniqueName: \"kubernetes.io/projected/6581e45a-6f33-45c5-a98f-c1fe77f6f703-kube-api-access-bj525\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745835 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745858 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745908 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.745918 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.747503 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.747701 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.748086 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.748162 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.754619 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-kube-api-access-7m9lp" (OuterVolumeSpecName: "kube-api-access-7m9lp") pod "f24fc3d7-4756-4abb-bcd7-2550349ef1ea" (UID: "f24fc3d7-4756-4abb-bcd7-2550349ef1ea"). InnerVolumeSpecName "kube-api-access-7m9lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.767966 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj525\" (UniqueName: \"kubernetes.io/projected/6581e45a-6f33-45c5-a98f-c1fe77f6f703-kube-api-access-bj525\") pod \"dnsmasq-dns-86db49b7ff-7g8pw\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.775265 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.790649 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.837539 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.846485 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kttd\" (UniqueName: \"kubernetes.io/projected/9f195fce-1d6b-4384-b733-c58ecd3fd16d-kube-api-access-9kttd\") pod \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.846602 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-dns-svc\") pod \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.846625 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-config\") pod \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\" (UID: \"9f195fce-1d6b-4384-b733-c58ecd3fd16d\") " Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.847018 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m9lp\" (UniqueName: \"kubernetes.io/projected/f24fc3d7-4756-4abb-bcd7-2550349ef1ea-kube-api-access-7m9lp\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.847384 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-config" (OuterVolumeSpecName: "config") pod "9f195fce-1d6b-4384-b733-c58ecd3fd16d" (UID: "9f195fce-1d6b-4384-b733-c58ecd3fd16d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.849282 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f195fce-1d6b-4384-b733-c58ecd3fd16d" (UID: "9f195fce-1d6b-4384-b733-c58ecd3fd16d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.851903 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f195fce-1d6b-4384-b733-c58ecd3fd16d-kube-api-access-9kttd" (OuterVolumeSpecName: "kube-api-access-9kttd") pod "9f195fce-1d6b-4384-b733-c58ecd3fd16d" (UID: "9f195fce-1d6b-4384-b733-c58ecd3fd16d"). InnerVolumeSpecName "kube-api-access-9kttd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.953647 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kttd\" (UniqueName: \"kubernetes.io/projected/9f195fce-1d6b-4384-b733-c58ecd3fd16d-kube-api-access-9kttd\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.953683 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:56 crc kubenswrapper[4758]: I1004 11:05:56.953692 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f195fce-1d6b-4384-b733-c58ecd3fd16d-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.082157 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6wnl7"] Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.229746 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-9cb54f9cb-xrmhr"] Oct 04 11:05:57 crc kubenswrapper[4758]: W1004 11:05:57.235957 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded03b984_16f3_46dd_a684_f381b6fbe89b.slice/crio-61012c4250628c6319d39a12747794f5172597029522054bf65f95de7018ce39 WatchSource:0}: Error finding container 61012c4250628c6319d39a12747794f5172597029522054bf65f95de7018ce39: Status 404 returned error can't find the container with id 61012c4250628c6319d39a12747794f5172597029522054bf65f95de7018ce39 Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.328188 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7g8pw"] Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.620446 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" event={"ID":"8001680b-56f4-47a4-94e2-891b1a56c2af","Type":"ContainerStarted","Data":"47db5cd6c77f35ddb38f70822c38fdca7c7a7ea866480471e7829ea6ccff85e0"} Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.621695 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" event={"ID":"6581e45a-6f33-45c5-a98f-c1fe77f6f703","Type":"ContainerStarted","Data":"a92741c435ed945c5c06f6d7dc4491c126eb7d4a322479159a82435d213e3b5f"} Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.623366 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" event={"ID":"ed03b984-16f3-46dd-a684-f381b6fbe89b","Type":"ContainerStarted","Data":"61012c4250628c6319d39a12747794f5172597029522054bf65f95de7018ce39"} Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.624528 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" event={"ID":"9f195fce-1d6b-4384-b733-c58ecd3fd16d","Type":"ContainerDied","Data":"455ea644323ecb3e0aeaec086ace4405a9b8012adb85d6893bb78c9d63086058"} Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.624552 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7bv59" Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.624602 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-hhg4h" Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.664475 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7bv59"] Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.671363 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7bv59"] Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.745751 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hhg4h"] Oct 04 11:05:57 crc kubenswrapper[4758]: I1004 11:05:57.751710 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-hhg4h"] Oct 04 11:05:58 crc kubenswrapper[4758]: I1004 11:05:58.632159 4758 generic.go:334] "Generic (PLEG): container finished" podID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerID="10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc" exitCode=0 Oct 04 11:05:58 crc kubenswrapper[4758]: I1004 11:05:58.632420 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" event={"ID":"6581e45a-6f33-45c5-a98f-c1fe77f6f703","Type":"ContainerDied","Data":"10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc"} Oct 04 11:05:58 crc kubenswrapper[4758]: I1004 11:05:58.635029 4758 generic.go:334] "Generic (PLEG): container finished" podID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerID="6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb" exitCode=0 Oct 04 11:05:58 crc kubenswrapper[4758]: I1004 11:05:58.635067 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" event={"ID":"8001680b-56f4-47a4-94e2-891b1a56c2af","Type":"ContainerDied","Data":"6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb"} Oct 04 11:05:58 crc kubenswrapper[4758]: I1004 11:05:58.670307 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.334455 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f195fce-1d6b-4384-b733-c58ecd3fd16d" path="/var/lib/kubelet/pods/9f195fce-1d6b-4384-b733-c58ecd3fd16d/volumes" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.335413 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f24fc3d7-4756-4abb-bcd7-2550349ef1ea" path="/var/lib/kubelet/pods/f24fc3d7-4756-4abb-bcd7-2550349ef1ea/volumes" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.642076 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" event={"ID":"ed03b984-16f3-46dd-a684-f381b6fbe89b","Type":"ContainerStarted","Data":"517d358addabffcc6bd47ff21cd966e1c9dcc7d2d0119fb96982399ac0c42fc9"} Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.642396 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.646368 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" event={"ID":"8001680b-56f4-47a4-94e2-891b1a56c2af","Type":"ContainerStarted","Data":"fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156"} Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.646476 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.648383 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" event={"ID":"6581e45a-6f33-45c5-a98f-c1fe77f6f703","Type":"ContainerStarted","Data":"f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49"} Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.648597 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.663757 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" podStartSLOduration=1.872745434 podStartE2EDuration="3.663738569s" podCreationTimestamp="2025-10-04 11:05:56 +0000 UTC" firstStartedPulling="2025-10-04 11:05:57.237940912 +0000 UTC m=+934.530591801" lastFinishedPulling="2025-10-04 11:05:59.028934047 +0000 UTC m=+936.321584936" observedRunningTime="2025-10-04 11:05:59.657711627 +0000 UTC m=+936.950362526" watchObservedRunningTime="2025-10-04 11:05:59.663738569 +0000 UTC m=+936.956389458" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.682064 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" podStartSLOduration=3.002564645 podStartE2EDuration="3.682044952s" podCreationTimestamp="2025-10-04 11:05:56 +0000 UTC" firstStartedPulling="2025-10-04 11:05:57.351543633 +0000 UTC m=+934.644194522" lastFinishedPulling="2025-10-04 11:05:58.03102395 +0000 UTC m=+935.323674829" observedRunningTime="2025-10-04 11:05:59.675420224 +0000 UTC m=+936.968071113" watchObservedRunningTime="2025-10-04 11:05:59.682044952 +0000 UTC m=+936.974695841" Oct 04 11:05:59 crc kubenswrapper[4758]: I1004 11:05:59.696389 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" podStartSLOduration=3.089165928 podStartE2EDuration="3.696369628s" podCreationTimestamp="2025-10-04 11:05:56 +0000 UTC" firstStartedPulling="2025-10-04 11:05:57.095043112 +0000 UTC m=+934.387693991" lastFinishedPulling="2025-10-04 11:05:57.702246802 +0000 UTC m=+934.994897691" observedRunningTime="2025-10-04 11:05:59.693375438 +0000 UTC m=+936.986026337" watchObservedRunningTime="2025-10-04 11:05:59.696369628 +0000 UTC m=+936.989020517" Oct 04 11:06:02 crc kubenswrapper[4758]: I1004 11:06:02.670606 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 11:06:02 crc kubenswrapper[4758]: I1004 11:06:02.670931 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 11:06:02 crc kubenswrapper[4758]: I1004 11:06:02.724050 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 11:06:03 crc kubenswrapper[4758]: I1004 11:06:03.003444 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 11:06:03 crc kubenswrapper[4758]: I1004 11:06:03.004971 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 11:06:03 crc kubenswrapper[4758]: I1004 11:06:03.050193 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 11:06:03 crc kubenswrapper[4758]: I1004 11:06:03.754344 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 11:06:03 crc kubenswrapper[4758]: I1004 11:06:03.792256 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.625794 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.729371 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6wnl7"] Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.729573 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerName="dnsmasq-dns" containerID="cri-o://fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156" gracePeriod=10 Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.731247 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.829256 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-4dncc"] Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.830468 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.845928 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4dncc"] Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.988139 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-config\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.988183 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-dns-svc\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.988211 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.988308 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc758\" (UniqueName: \"kubernetes.io/projected/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-kube-api-access-tc758\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:04 crc kubenswrapper[4758]: I1004 11:06:04.988345 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.090000 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.090404 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-config\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.090654 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-dns-svc\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.090697 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.090783 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc758\" (UniqueName: \"kubernetes.io/projected/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-kube-api-access-tc758\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.091092 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.091955 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-dns-svc\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.091972 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-config\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.092325 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.115828 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc758\" (UniqueName: \"kubernetes.io/projected/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-kube-api-access-tc758\") pod \"dnsmasq-dns-698758b865-4dncc\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.164968 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.420814 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.496788 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-ovsdbserver-nb\") pod \"8001680b-56f4-47a4-94e2-891b1a56c2af\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.496863 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-config\") pod \"8001680b-56f4-47a4-94e2-891b1a56c2af\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.496976 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-dns-svc\") pod \"8001680b-56f4-47a4-94e2-891b1a56c2af\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.496995 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xbl2\" (UniqueName: \"kubernetes.io/projected/8001680b-56f4-47a4-94e2-891b1a56c2af-kube-api-access-2xbl2\") pod \"8001680b-56f4-47a4-94e2-891b1a56c2af\" (UID: \"8001680b-56f4-47a4-94e2-891b1a56c2af\") " Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.500990 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8001680b-56f4-47a4-94e2-891b1a56c2af-kube-api-access-2xbl2" (OuterVolumeSpecName: "kube-api-access-2xbl2") pod "8001680b-56f4-47a4-94e2-891b1a56c2af" (UID: "8001680b-56f4-47a4-94e2-891b1a56c2af"). InnerVolumeSpecName "kube-api-access-2xbl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.535198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-config" (OuterVolumeSpecName: "config") pod "8001680b-56f4-47a4-94e2-891b1a56c2af" (UID: "8001680b-56f4-47a4-94e2-891b1a56c2af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.537733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8001680b-56f4-47a4-94e2-891b1a56c2af" (UID: "8001680b-56f4-47a4-94e2-891b1a56c2af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.543240 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8001680b-56f4-47a4-94e2-891b1a56c2af" (UID: "8001680b-56f4-47a4-94e2-891b1a56c2af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.599135 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.599164 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xbl2\" (UniqueName: \"kubernetes.io/projected/8001680b-56f4-47a4-94e2-891b1a56c2af-kube-api-access-2xbl2\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.599174 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.599184 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8001680b-56f4-47a4-94e2-891b1a56c2af-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.693960 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4dncc"] Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.694156 4758 generic.go:334] "Generic (PLEG): container finished" podID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerID="fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156" exitCode=0 Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.694927 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.695723 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" event={"ID":"8001680b-56f4-47a4-94e2-891b1a56c2af","Type":"ContainerDied","Data":"fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156"} Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.695761 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6wnl7" event={"ID":"8001680b-56f4-47a4-94e2-891b1a56c2af","Type":"ContainerDied","Data":"47db5cd6c77f35ddb38f70822c38fdca7c7a7ea866480471e7829ea6ccff85e0"} Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.695778 4758 scope.go:117] "RemoveContainer" containerID="fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.724637 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6wnl7"] Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.725087 4758 scope.go:117] "RemoveContainer" containerID="6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.730922 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6wnl7"] Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.745066 4758 scope.go:117] "RemoveContainer" containerID="fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156" Oct 04 11:06:05 crc kubenswrapper[4758]: E1004 11:06:05.746498 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156\": container with ID starting with fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156 not found: ID does not exist" containerID="fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.746589 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156"} err="failed to get container status \"fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156\": rpc error: code = NotFound desc = could not find container \"fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156\": container with ID starting with fd0208ddbc193d1df4456d9927adc2f752005b6f94a5299514129fa62003b156 not found: ID does not exist" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.746658 4758 scope.go:117] "RemoveContainer" containerID="6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb" Oct 04 11:06:05 crc kubenswrapper[4758]: E1004 11:06:05.747139 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb\": container with ID starting with 6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb not found: ID does not exist" containerID="6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.747184 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb"} err="failed to get container status \"6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb\": rpc error: code = NotFound desc = could not find container \"6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb\": container with ID starting with 6170f5b82813214fec714b61aa213b3981d8425a5ebed9fe93a6dabdae9f10eb not found: ID does not exist" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.748521 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 11:06:05 crc kubenswrapper[4758]: E1004 11:06:05.748975 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerName="dnsmasq-dns" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.749054 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerName="dnsmasq-dns" Oct 04 11:06:05 crc kubenswrapper[4758]: E1004 11:06:05.749145 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerName="init" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.749853 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerName="init" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.750072 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" containerName="dnsmasq-dns" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.756220 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.762529 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lhkj2" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.762700 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.763525 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.763711 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.768215 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.903134 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d3ed3476-2568-4c27-ac70-87d732389beb-cache\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.903856 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9x4q\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-kube-api-access-f9x4q\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.903964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.904055 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:05 crc kubenswrapper[4758]: I1004 11:06:05.904204 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d3ed3476-2568-4c27-ac70-87d732389beb-lock\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.005510 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.005569 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d3ed3476-2568-4c27-ac70-87d732389beb-lock\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.005636 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d3ed3476-2568-4c27-ac70-87d732389beb-cache\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.005658 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9x4q\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-kube-api-access-f9x4q\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.005699 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: E1004 11:06:06.005843 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 11:06:06 crc kubenswrapper[4758]: E1004 11:06:06.005857 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 11:06:06 crc kubenswrapper[4758]: E1004 11:06:06.005904 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift podName:d3ed3476-2568-4c27-ac70-87d732389beb nodeName:}" failed. No retries permitted until 2025-10-04 11:06:06.505885904 +0000 UTC m=+943.798536793 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift") pod "swift-storage-0" (UID: "d3ed3476-2568-4c27-ac70-87d732389beb") : configmap "swift-ring-files" not found Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.006536 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.006770 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d3ed3476-2568-4c27-ac70-87d732389beb-cache\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.006831 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d3ed3476-2568-4c27-ac70-87d732389beb-lock\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.027769 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9x4q\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-kube-api-access-f9x4q\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.033252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.399893 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-b9wqh"] Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.403391 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.409255 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b9wqh"] Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.411138 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.411416 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.411602 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514328 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514399 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a127cd92-1013-49e5-b5ec-832428b0b113-etc-swift\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514425 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-swiftconf\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514446 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-dispersionconf\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514535 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-scripts\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: E1004 11:06:06.514534 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 11:06:06 crc kubenswrapper[4758]: E1004 11:06:06.514566 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514554 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-ring-data-devices\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: E1004 11:06:06.514615 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift podName:d3ed3476-2568-4c27-ac70-87d732389beb nodeName:}" failed. No retries permitted until 2025-10-04 11:06:07.514599371 +0000 UTC m=+944.807250260 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift") pod "swift-storage-0" (UID: "d3ed3476-2568-4c27-ac70-87d732389beb") : configmap "swift-ring-files" not found Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514771 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9sqd\" (UniqueName: \"kubernetes.io/projected/a127cd92-1013-49e5-b5ec-832428b0b113-kube-api-access-x9sqd\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.514827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-combined-ca-bundle\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.615631 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9sqd\" (UniqueName: \"kubernetes.io/projected/a127cd92-1013-49e5-b5ec-832428b0b113-kube-api-access-x9sqd\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.616227 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-combined-ca-bundle\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.616374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a127cd92-1013-49e5-b5ec-832428b0b113-etc-swift\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.616497 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-swiftconf\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.616604 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-dispersionconf\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.616749 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-scripts\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.617523 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-ring-data-devices\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.617561 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-scripts\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.616818 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a127cd92-1013-49e5-b5ec-832428b0b113-etc-swift\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.617994 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-ring-data-devices\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.620488 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-combined-ca-bundle\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.620571 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-swiftconf\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.630512 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-dispersionconf\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.635426 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9sqd\" (UniqueName: \"kubernetes.io/projected/a127cd92-1013-49e5-b5ec-832428b0b113-kube-api-access-x9sqd\") pod \"swift-ring-rebalance-b9wqh\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.701337 4758 generic.go:334] "Generic (PLEG): container finished" podID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerID="88e13b31bc6f1c35db9adbbd675c0e2c908a20cf19cce67277c3bdd1275f059b" exitCode=0 Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.701399 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4dncc" event={"ID":"9f9584b6-569d-4e7a-bfd1-df47048bd3aa","Type":"ContainerDied","Data":"88e13b31bc6f1c35db9adbbd675c0e2c908a20cf19cce67277c3bdd1275f059b"} Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.701424 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4dncc" event={"ID":"9f9584b6-569d-4e7a-bfd1-df47048bd3aa","Type":"ContainerStarted","Data":"c28fd65c5dafcc86ade061d0837115c8e24e9f69b521df9d168640a598016c43"} Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.724277 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:06 crc kubenswrapper[4758]: I1004 11:06:06.840398 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.183830 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b9wqh"] Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.335270 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8001680b-56f4-47a4-94e2-891b1a56c2af" path="/var/lib/kubelet/pods/8001680b-56f4-47a4-94e2-891b1a56c2af/volumes" Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.539059 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:07 crc kubenswrapper[4758]: E1004 11:06:07.539287 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 11:06:07 crc kubenswrapper[4758]: E1004 11:06:07.540257 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 11:06:07 crc kubenswrapper[4758]: E1004 11:06:07.540319 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift podName:d3ed3476-2568-4c27-ac70-87d732389beb nodeName:}" failed. No retries permitted until 2025-10-04 11:06:09.540299045 +0000 UTC m=+946.832949954 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift") pod "swift-storage-0" (UID: "d3ed3476-2568-4c27-ac70-87d732389beb") : configmap "swift-ring-files" not found Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.716726 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4dncc" event={"ID":"9f9584b6-569d-4e7a-bfd1-df47048bd3aa","Type":"ContainerStarted","Data":"22036001acb4ea1abcfa323d47d4df958c65d491a024808769b6ad1ba44f5294"} Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.717663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.720869 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9wqh" event={"ID":"a127cd92-1013-49e5-b5ec-832428b0b113","Type":"ContainerStarted","Data":"d705309ce1251d4e430875a84ae56431cd5078b14d1a4ad9005626169ee98378"} Oct 04 11:06:07 crc kubenswrapper[4758]: I1004 11:06:07.750830 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-4dncc" podStartSLOduration=3.750804117 podStartE2EDuration="3.750804117s" podCreationTimestamp="2025-10-04 11:06:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:06:07.748934886 +0000 UTC m=+945.041585805" watchObservedRunningTime="2025-10-04 11:06:07.750804117 +0000 UTC m=+945.043455036" Oct 04 11:06:08 crc kubenswrapper[4758]: I1004 11:06:08.923551 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-68x7w"] Oct 04 11:06:08 crc kubenswrapper[4758]: I1004 11:06:08.925361 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-68x7w" Oct 04 11:06:08 crc kubenswrapper[4758]: I1004 11:06:08.935608 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-68x7w"] Oct 04 11:06:09 crc kubenswrapper[4758]: I1004 11:06:09.072975 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ggh\" (UniqueName: \"kubernetes.io/projected/66b43b9d-d8b6-45e8-83ae-432775163641-kube-api-access-r8ggh\") pod \"glance-db-create-68x7w\" (UID: \"66b43b9d-d8b6-45e8-83ae-432775163641\") " pod="openstack/glance-db-create-68x7w" Oct 04 11:06:09 crc kubenswrapper[4758]: I1004 11:06:09.174707 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ggh\" (UniqueName: \"kubernetes.io/projected/66b43b9d-d8b6-45e8-83ae-432775163641-kube-api-access-r8ggh\") pod \"glance-db-create-68x7w\" (UID: \"66b43b9d-d8b6-45e8-83ae-432775163641\") " pod="openstack/glance-db-create-68x7w" Oct 04 11:06:09 crc kubenswrapper[4758]: I1004 11:06:09.197221 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ggh\" (UniqueName: \"kubernetes.io/projected/66b43b9d-d8b6-45e8-83ae-432775163641-kube-api-access-r8ggh\") pod \"glance-db-create-68x7w\" (UID: \"66b43b9d-d8b6-45e8-83ae-432775163641\") " pod="openstack/glance-db-create-68x7w" Oct 04 11:06:09 crc kubenswrapper[4758]: I1004 11:06:09.281393 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-68x7w" Oct 04 11:06:09 crc kubenswrapper[4758]: I1004 11:06:09.581208 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:09 crc kubenswrapper[4758]: E1004 11:06:09.581369 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 11:06:09 crc kubenswrapper[4758]: E1004 11:06:09.581382 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 11:06:09 crc kubenswrapper[4758]: E1004 11:06:09.581421 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift podName:d3ed3476-2568-4c27-ac70-87d732389beb nodeName:}" failed. No retries permitted until 2025-10-04 11:06:13.581407698 +0000 UTC m=+950.874058587 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift") pod "swift-storage-0" (UID: "d3ed3476-2568-4c27-ac70-87d732389beb") : configmap "swift-ring-files" not found Oct 04 11:06:11 crc kubenswrapper[4758]: I1004 11:06:11.374508 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-68x7w"] Oct 04 11:06:11 crc kubenswrapper[4758]: I1004 11:06:11.855081 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-9cb54f9cb-xrmhr" Oct 04 11:06:12 crc kubenswrapper[4758]: W1004 11:06:12.052312 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66b43b9d_d8b6_45e8_83ae_432775163641.slice/crio-e5f805b92c85200f9f8496a19ec7e324e1e2ae5e08e88fe5494aa5131992ad66 WatchSource:0}: Error finding container e5f805b92c85200f9f8496a19ec7e324e1e2ae5e08e88fe5494aa5131992ad66: Status 404 returned error can't find the container with id e5f805b92c85200f9f8496a19ec7e324e1e2ae5e08e88fe5494aa5131992ad66 Oct 04 11:06:12 crc kubenswrapper[4758]: I1004 11:06:12.784269 4758 generic.go:334] "Generic (PLEG): container finished" podID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerID="a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17" exitCode=0 Oct 04 11:06:12 crc kubenswrapper[4758]: I1004 11:06:12.784605 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"57d6ac32-db38-4443-9c0c-13c66db2451b","Type":"ContainerDied","Data":"a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17"} Oct 04 11:06:12 crc kubenswrapper[4758]: I1004 11:06:12.792228 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-68x7w" event={"ID":"66b43b9d-d8b6-45e8-83ae-432775163641","Type":"ContainerStarted","Data":"e5f805b92c85200f9f8496a19ec7e324e1e2ae5e08e88fe5494aa5131992ad66"} Oct 04 11:06:12 crc kubenswrapper[4758]: I1004 11:06:12.817513 4758 generic.go:334] "Generic (PLEG): container finished" podID="acd7d674-14dc-4292-a4b1-5175299225e0" containerID="fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4" exitCode=0 Oct 04 11:06:12 crc kubenswrapper[4758]: I1004 11:06:12.817554 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"acd7d674-14dc-4292-a4b1-5175299225e0","Type":"ContainerDied","Data":"fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4"} Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.239141 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-z6vwk"] Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.240233 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.251594 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-z6vwk"] Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.342385 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz52z\" (UniqueName: \"kubernetes.io/projected/6f88e571-0f7b-4740-a4e2-3ce25c94bac9-kube-api-access-fz52z\") pod \"keystone-db-create-z6vwk\" (UID: \"6f88e571-0f7b-4740-a4e2-3ce25c94bac9\") " pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.443839 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz52z\" (UniqueName: \"kubernetes.io/projected/6f88e571-0f7b-4740-a4e2-3ce25c94bac9-kube-api-access-fz52z\") pod \"keystone-db-create-z6vwk\" (UID: \"6f88e571-0f7b-4740-a4e2-3ce25c94bac9\") " pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.471149 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz52z\" (UniqueName: \"kubernetes.io/projected/6f88e571-0f7b-4740-a4e2-3ce25c94bac9-kube-api-access-fz52z\") pod \"keystone-db-create-z6vwk\" (UID: \"6f88e571-0f7b-4740-a4e2-3ce25c94bac9\") " pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.542806 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ktlf7"] Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.544143 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.558246 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ktlf7"] Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.559038 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.646273 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gwc2\" (UniqueName: \"kubernetes.io/projected/5b59f374-8731-40e2-b34d-4874ba393094-kube-api-access-7gwc2\") pod \"placement-db-create-ktlf7\" (UID: \"5b59f374-8731-40e2-b34d-4874ba393094\") " pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.646337 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:13 crc kubenswrapper[4758]: E1004 11:06:13.646490 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 11:06:13 crc kubenswrapper[4758]: E1004 11:06:13.646505 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 11:06:13 crc kubenswrapper[4758]: E1004 11:06:13.646545 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift podName:d3ed3476-2568-4c27-ac70-87d732389beb nodeName:}" failed. No retries permitted until 2025-10-04 11:06:21.646529344 +0000 UTC m=+958.939180233 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift") pod "swift-storage-0" (UID: "d3ed3476-2568-4c27-ac70-87d732389beb") : configmap "swift-ring-files" not found Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.748259 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gwc2\" (UniqueName: \"kubernetes.io/projected/5b59f374-8731-40e2-b34d-4874ba393094-kube-api-access-7gwc2\") pod \"placement-db-create-ktlf7\" (UID: \"5b59f374-8731-40e2-b34d-4874ba393094\") " pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.765889 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gwc2\" (UniqueName: \"kubernetes.io/projected/5b59f374-8731-40e2-b34d-4874ba393094-kube-api-access-7gwc2\") pod \"placement-db-create-ktlf7\" (UID: \"5b59f374-8731-40e2-b34d-4874ba393094\") " pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:13 crc kubenswrapper[4758]: I1004 11:06:13.875258 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.015829 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-z6vwk"] Oct 04 11:06:14 crc kubenswrapper[4758]: W1004 11:06:14.026223 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f88e571_0f7b_4740_a4e2_3ce25c94bac9.slice/crio-7c1ab4f51ce40bd20ce44c550d94d80e2b1ae811bcbcb1206c8ecf7cb6e517b8 WatchSource:0}: Error finding container 7c1ab4f51ce40bd20ce44c550d94d80e2b1ae811bcbcb1206c8ecf7cb6e517b8: Status 404 returned error can't find the container with id 7c1ab4f51ce40bd20ce44c550d94d80e2b1ae811bcbcb1206c8ecf7cb6e517b8 Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.304950 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ktlf7"] Oct 04 11:06:14 crc kubenswrapper[4758]: W1004 11:06:14.312900 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b59f374_8731_40e2_b34d_4874ba393094.slice/crio-b08dbd6f72931b12d43bd16e6487e9760970343f1cf849ab39497bb153bf60f8 WatchSource:0}: Error finding container b08dbd6f72931b12d43bd16e6487e9760970343f1cf849ab39497bb153bf60f8: Status 404 returned error can't find the container with id b08dbd6f72931b12d43bd16e6487e9760970343f1cf849ab39497bb153bf60f8 Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.834717 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"57d6ac32-db38-4443-9c0c-13c66db2451b","Type":"ContainerStarted","Data":"77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.835410 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.838964 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktlf7" event={"ID":"5b59f374-8731-40e2-b34d-4874ba393094","Type":"ContainerStarted","Data":"dbcfb67bfdf15825765f735285dd714869298a659ccc5fd8587c1ca1ec10827b"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.839011 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktlf7" event={"ID":"5b59f374-8731-40e2-b34d-4874ba393094","Type":"ContainerStarted","Data":"b08dbd6f72931b12d43bd16e6487e9760970343f1cf849ab39497bb153bf60f8"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.841745 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-68x7w" event={"ID":"66b43b9d-d8b6-45e8-83ae-432775163641","Type":"ContainerStarted","Data":"f0c1585415240feabf438e7f5a3d3c0965720f7d757687072f27536d4f50308d"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.843973 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z6vwk" event={"ID":"6f88e571-0f7b-4740-a4e2-3ce25c94bac9","Type":"ContainerStarted","Data":"9b164ac84dbd8ab507cbdea2656f78d8cbe32079002f4b0075d67b399025af07"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.844161 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z6vwk" event={"ID":"6f88e571-0f7b-4740-a4e2-3ce25c94bac9","Type":"ContainerStarted","Data":"7c1ab4f51ce40bd20ce44c550d94d80e2b1ae811bcbcb1206c8ecf7cb6e517b8"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.847557 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"acd7d674-14dc-4292-a4b1-5175299225e0","Type":"ContainerStarted","Data":"765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.848423 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.850501 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9wqh" event={"ID":"a127cd92-1013-49e5-b5ec-832428b0b113","Type":"ContainerStarted","Data":"e28f230d38105200fbf53e562373965539e3d9f89c2ae3ee61bd6eaf004ff8ef"} Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.863284 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=34.859423126 podStartE2EDuration="55.863264065s" podCreationTimestamp="2025-10-04 11:05:19 +0000 UTC" firstStartedPulling="2025-10-04 11:05:20.979205475 +0000 UTC m=+898.271856364" lastFinishedPulling="2025-10-04 11:05:41.983046424 +0000 UTC m=+919.275697303" observedRunningTime="2025-10-04 11:06:14.860216313 +0000 UTC m=+952.152867192" watchObservedRunningTime="2025-10-04 11:06:14.863264065 +0000 UTC m=+952.155914954" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.879231 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-68x7w" podStartSLOduration=6.879217165 podStartE2EDuration="6.879217165s" podCreationTimestamp="2025-10-04 11:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:06:14.875018502 +0000 UTC m=+952.167669391" watchObservedRunningTime="2025-10-04 11:06:14.879217165 +0000 UTC m=+952.171868054" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.896570 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-z6vwk" podStartSLOduration=1.896552932 podStartE2EDuration="1.896552932s" podCreationTimestamp="2025-10-04 11:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:06:14.892309118 +0000 UTC m=+952.184960007" watchObservedRunningTime="2025-10-04 11:06:14.896552932 +0000 UTC m=+952.189203821" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.939316 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.828199475 podStartE2EDuration="56.939299384s" podCreationTimestamp="2025-10-04 11:05:18 +0000 UTC" firstStartedPulling="2025-10-04 11:05:20.785672791 +0000 UTC m=+898.078323680" lastFinishedPulling="2025-10-04 11:05:41.89677269 +0000 UTC m=+919.189423589" observedRunningTime="2025-10-04 11:06:14.934348031 +0000 UTC m=+952.226998940" watchObservedRunningTime="2025-10-04 11:06:14.939299384 +0000 UTC m=+952.231950273" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.975931 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-b9wqh" podStartSLOduration=3.884629687 podStartE2EDuration="8.97590261s" podCreationTimestamp="2025-10-04 11:06:06 +0000 UTC" firstStartedPulling="2025-10-04 11:06:07.186256586 +0000 UTC m=+944.478907475" lastFinishedPulling="2025-10-04 11:06:12.277529499 +0000 UTC m=+949.570180398" observedRunningTime="2025-10-04 11:06:14.970812243 +0000 UTC m=+952.263463132" watchObservedRunningTime="2025-10-04 11:06:14.97590261 +0000 UTC m=+952.268553499" Oct 04 11:06:14 crc kubenswrapper[4758]: I1004 11:06:14.986628 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-ktlf7" podStartSLOduration=1.986610499 podStartE2EDuration="1.986610499s" podCreationTimestamp="2025-10-04 11:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:06:14.983536006 +0000 UTC m=+952.276186895" watchObservedRunningTime="2025-10-04 11:06:14.986610499 +0000 UTC m=+952.279261388" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.167443 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.223838 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7g8pw"] Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.224156 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerName="dnsmasq-dns" containerID="cri-o://f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49" gracePeriod=10 Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.782538 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.859216 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b59f374-8731-40e2-b34d-4874ba393094" containerID="dbcfb67bfdf15825765f735285dd714869298a659ccc5fd8587c1ca1ec10827b" exitCode=0 Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.859693 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktlf7" event={"ID":"5b59f374-8731-40e2-b34d-4874ba393094","Type":"ContainerDied","Data":"dbcfb67bfdf15825765f735285dd714869298a659ccc5fd8587c1ca1ec10827b"} Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.861333 4758 generic.go:334] "Generic (PLEG): container finished" podID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerID="f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49" exitCode=0 Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.861366 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" event={"ID":"6581e45a-6f33-45c5-a98f-c1fe77f6f703","Type":"ContainerDied","Data":"f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49"} Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.861384 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" event={"ID":"6581e45a-6f33-45c5-a98f-c1fe77f6f703","Type":"ContainerDied","Data":"a92741c435ed945c5c06f6d7dc4491c126eb7d4a322479159a82435d213e3b5f"} Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.861401 4758 scope.go:117] "RemoveContainer" containerID="f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.861497 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-7g8pw" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.863800 4758 generic.go:334] "Generic (PLEG): container finished" podID="66b43b9d-d8b6-45e8-83ae-432775163641" containerID="f0c1585415240feabf438e7f5a3d3c0965720f7d757687072f27536d4f50308d" exitCode=0 Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.863900 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-68x7w" event={"ID":"66b43b9d-d8b6-45e8-83ae-432775163641","Type":"ContainerDied","Data":"f0c1585415240feabf438e7f5a3d3c0965720f7d757687072f27536d4f50308d"} Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.870721 4758 generic.go:334] "Generic (PLEG): container finished" podID="6f88e571-0f7b-4740-a4e2-3ce25c94bac9" containerID="9b164ac84dbd8ab507cbdea2656f78d8cbe32079002f4b0075d67b399025af07" exitCode=0 Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.870862 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z6vwk" event={"ID":"6f88e571-0f7b-4740-a4e2-3ce25c94bac9","Type":"ContainerDied","Data":"9b164ac84dbd8ab507cbdea2656f78d8cbe32079002f4b0075d67b399025af07"} Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.882446 4758 scope.go:117] "RemoveContainer" containerID="10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.887576 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-dns-svc\") pod \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.887633 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj525\" (UniqueName: \"kubernetes.io/projected/6581e45a-6f33-45c5-a98f-c1fe77f6f703-kube-api-access-bj525\") pod \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.887654 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-sb\") pod \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.887722 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config\") pod \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.887831 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-nb\") pod \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.898750 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6581e45a-6f33-45c5-a98f-c1fe77f6f703-kube-api-access-bj525" (OuterVolumeSpecName: "kube-api-access-bj525") pod "6581e45a-6f33-45c5-a98f-c1fe77f6f703" (UID: "6581e45a-6f33-45c5-a98f-c1fe77f6f703"). InnerVolumeSpecName "kube-api-access-bj525". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.965311 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6581e45a-6f33-45c5-a98f-c1fe77f6f703" (UID: "6581e45a-6f33-45c5-a98f-c1fe77f6f703"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.970599 4758 scope.go:117] "RemoveContainer" containerID="f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49" Oct 04 11:06:15 crc kubenswrapper[4758]: E1004 11:06:15.974539 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49\": container with ID starting with f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49 not found: ID does not exist" containerID="f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.974587 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49"} err="failed to get container status \"f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49\": rpc error: code = NotFound desc = could not find container \"f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49\": container with ID starting with f947b41792625383a15e402e0845555e01391699919b2294713dc74867ff7b49 not found: ID does not exist" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.974606 4758 scope.go:117] "RemoveContainer" containerID="10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc" Oct 04 11:06:15 crc kubenswrapper[4758]: E1004 11:06:15.975744 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc\": container with ID starting with 10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc not found: ID does not exist" containerID="10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.975777 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc"} err="failed to get container status \"10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc\": rpc error: code = NotFound desc = could not find container \"10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc\": container with ID starting with 10a3aa5234cbcc70063928c33ecb6ba0bed19af94f93d5ea51e102271a9aafbc not found: ID does not exist" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.989166 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config" (OuterVolumeSpecName: "config") pod "6581e45a-6f33-45c5-a98f-c1fe77f6f703" (UID: "6581e45a-6f33-45c5-a98f-c1fe77f6f703"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.989483 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config\") pod \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\" (UID: \"6581e45a-6f33-45c5-a98f-c1fe77f6f703\") " Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.989815 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj525\" (UniqueName: \"kubernetes.io/projected/6581e45a-6f33-45c5-a98f-c1fe77f6f703-kube-api-access-bj525\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.989828 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:15 crc kubenswrapper[4758]: W1004 11:06:15.989887 4758 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/6581e45a-6f33-45c5-a98f-c1fe77f6f703/volumes/kubernetes.io~configmap/config Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.989897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config" (OuterVolumeSpecName: "config") pod "6581e45a-6f33-45c5-a98f-c1fe77f6f703" (UID: "6581e45a-6f33-45c5-a98f-c1fe77f6f703"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:15 crc kubenswrapper[4758]: I1004 11:06:15.995087 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6581e45a-6f33-45c5-a98f-c1fe77f6f703" (UID: "6581e45a-6f33-45c5-a98f-c1fe77f6f703"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:16 crc kubenswrapper[4758]: I1004 11:06:16.015627 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6581e45a-6f33-45c5-a98f-c1fe77f6f703" (UID: "6581e45a-6f33-45c5-a98f-c1fe77f6f703"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:16 crc kubenswrapper[4758]: I1004 11:06:16.091411 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:16 crc kubenswrapper[4758]: I1004 11:06:16.091451 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:16 crc kubenswrapper[4758]: I1004 11:06:16.091464 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6581e45a-6f33-45c5-a98f-c1fe77f6f703-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:16 crc kubenswrapper[4758]: I1004 11:06:16.196356 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7g8pw"] Oct 04 11:06:16 crc kubenswrapper[4758]: I1004 11:06:16.203175 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-7g8pw"] Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.339153 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" path="/var/lib/kubelet/pods/6581e45a-6f33-45c5-a98f-c1fe77f6f703/volumes" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.344669 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.511678 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gwc2\" (UniqueName: \"kubernetes.io/projected/5b59f374-8731-40e2-b34d-4874ba393094-kube-api-access-7gwc2\") pod \"5b59f374-8731-40e2-b34d-4874ba393094\" (UID: \"5b59f374-8731-40e2-b34d-4874ba393094\") " Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.517517 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b59f374-8731-40e2-b34d-4874ba393094-kube-api-access-7gwc2" (OuterVolumeSpecName: "kube-api-access-7gwc2") pod "5b59f374-8731-40e2-b34d-4874ba393094" (UID: "5b59f374-8731-40e2-b34d-4874ba393094"). InnerVolumeSpecName "kube-api-access-7gwc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.560303 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.563886 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-68x7w" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.614000 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gwc2\" (UniqueName: \"kubernetes.io/projected/5b59f374-8731-40e2-b34d-4874ba393094-kube-api-access-7gwc2\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.714561 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8ggh\" (UniqueName: \"kubernetes.io/projected/66b43b9d-d8b6-45e8-83ae-432775163641-kube-api-access-r8ggh\") pod \"66b43b9d-d8b6-45e8-83ae-432775163641\" (UID: \"66b43b9d-d8b6-45e8-83ae-432775163641\") " Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.714667 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz52z\" (UniqueName: \"kubernetes.io/projected/6f88e571-0f7b-4740-a4e2-3ce25c94bac9-kube-api-access-fz52z\") pod \"6f88e571-0f7b-4740-a4e2-3ce25c94bac9\" (UID: \"6f88e571-0f7b-4740-a4e2-3ce25c94bac9\") " Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.720227 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b43b9d-d8b6-45e8-83ae-432775163641-kube-api-access-r8ggh" (OuterVolumeSpecName: "kube-api-access-r8ggh") pod "66b43b9d-d8b6-45e8-83ae-432775163641" (UID: "66b43b9d-d8b6-45e8-83ae-432775163641"). InnerVolumeSpecName "kube-api-access-r8ggh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.727898 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f88e571-0f7b-4740-a4e2-3ce25c94bac9-kube-api-access-fz52z" (OuterVolumeSpecName: "kube-api-access-fz52z") pod "6f88e571-0f7b-4740-a4e2-3ce25c94bac9" (UID: "6f88e571-0f7b-4740-a4e2-3ce25c94bac9"). InnerVolumeSpecName "kube-api-access-fz52z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.816988 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8ggh\" (UniqueName: \"kubernetes.io/projected/66b43b9d-d8b6-45e8-83ae-432775163641-kube-api-access-r8ggh\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.817022 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz52z\" (UniqueName: \"kubernetes.io/projected/6f88e571-0f7b-4740-a4e2-3ce25c94bac9-kube-api-access-fz52z\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.895248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ktlf7" event={"ID":"5b59f374-8731-40e2-b34d-4874ba393094","Type":"ContainerDied","Data":"b08dbd6f72931b12d43bd16e6487e9760970343f1cf849ab39497bb153bf60f8"} Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.895281 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ktlf7" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.895296 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b08dbd6f72931b12d43bd16e6487e9760970343f1cf849ab39497bb153bf60f8" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.896539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-68x7w" event={"ID":"66b43b9d-d8b6-45e8-83ae-432775163641","Type":"ContainerDied","Data":"e5f805b92c85200f9f8496a19ec7e324e1e2ae5e08e88fe5494aa5131992ad66"} Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.896561 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5f805b92c85200f9f8496a19ec7e324e1e2ae5e08e88fe5494aa5131992ad66" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.896627 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-68x7w" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.899124 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z6vwk" event={"ID":"6f88e571-0f7b-4740-a4e2-3ce25c94bac9","Type":"ContainerDied","Data":"7c1ab4f51ce40bd20ce44c550d94d80e2b1ae811bcbcb1206c8ecf7cb6e517b8"} Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.899332 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c1ab4f51ce40bd20ce44c550d94d80e2b1ae811bcbcb1206c8ecf7cb6e517b8" Oct 04 11:06:17 crc kubenswrapper[4758]: I1004 11:06:17.899306 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z6vwk" Oct 04 11:06:21 crc kubenswrapper[4758]: I1004 11:06:21.688111 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:21 crc kubenswrapper[4758]: E1004 11:06:21.688476 4758 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 11:06:21 crc kubenswrapper[4758]: E1004 11:06:21.688779 4758 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 11:06:21 crc kubenswrapper[4758]: E1004 11:06:21.688838 4758 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift podName:d3ed3476-2568-4c27-ac70-87d732389beb nodeName:}" failed. No retries permitted until 2025-10-04 11:06:37.688819534 +0000 UTC m=+974.981470433 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift") pod "swift-storage-0" (UID: "d3ed3476-2568-4c27-ac70-87d732389beb") : configmap "swift-ring-files" not found Oct 04 11:06:22 crc kubenswrapper[4758]: I1004 11:06:22.932439 4758 generic.go:334] "Generic (PLEG): container finished" podID="a127cd92-1013-49e5-b5ec-832428b0b113" containerID="e28f230d38105200fbf53e562373965539e3d9f89c2ae3ee61bd6eaf004ff8ef" exitCode=0 Oct 04 11:06:22 crc kubenswrapper[4758]: I1004 11:06:22.933014 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9wqh" event={"ID":"a127cd92-1013-49e5-b5ec-832428b0b113","Type":"ContainerDied","Data":"e28f230d38105200fbf53e562373965539e3d9f89c2ae3ee61bd6eaf004ff8ef"} Oct 04 11:06:22 crc kubenswrapper[4758]: I1004 11:06:22.987074 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gpsfr" podUID="39bf6bd9-f043-4485-8980-846341f31b4e" containerName="ovn-controller" probeResult="failure" output=< Oct 04 11:06:22 crc kubenswrapper[4758]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 11:06:22 crc kubenswrapper[4758]: > Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.058663 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.071176 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7fwv4" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.383122 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-80cb-account-create-qcgm2"] Oct 04 11:06:23 crc kubenswrapper[4758]: E1004 11:06:23.383823 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b43b9d-d8b6-45e8-83ae-432775163641" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.383840 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b43b9d-d8b6-45e8-83ae-432775163641" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: E1004 11:06:23.383851 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerName="init" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.383858 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerName="init" Oct 04 11:06:23 crc kubenswrapper[4758]: E1004 11:06:23.383870 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f88e571-0f7b-4740-a4e2-3ce25c94bac9" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.383877 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f88e571-0f7b-4740-a4e2-3ce25c94bac9" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: E1004 11:06:23.383886 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b59f374-8731-40e2-b34d-4874ba393094" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.383891 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b59f374-8731-40e2-b34d-4874ba393094" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: E1004 11:06:23.383906 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerName="dnsmasq-dns" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.383912 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerName="dnsmasq-dns" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.384065 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f88e571-0f7b-4740-a4e2-3ce25c94bac9" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.384075 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6581e45a-6f33-45c5-a98f-c1fe77f6f703" containerName="dnsmasq-dns" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.384089 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b43b9d-d8b6-45e8-83ae-432775163641" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.384116 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b59f374-8731-40e2-b34d-4874ba393094" containerName="mariadb-database-create" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.384659 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.388350 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.401341 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-80cb-account-create-qcgm2"] Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.462761 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gpsfr-config-tj2zs"] Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.463678 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.525528 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m6w4\" (UniqueName: \"kubernetes.io/projected/8849e0bd-1093-4779-a8ae-16ded0d8622e-kube-api-access-7m6w4\") pod \"keystone-80cb-account-create-qcgm2\" (UID: \"8849e0bd-1093-4779-a8ae-16ded0d8622e\") " pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.557633 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gpsfr-config-tj2zs"] Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.627382 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jflb\" (UniqueName: \"kubernetes.io/projected/48dd72b8-dbca-45bc-b450-eb05338f8fc0-kube-api-access-9jflb\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.627710 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m6w4\" (UniqueName: \"kubernetes.io/projected/8849e0bd-1093-4779-a8ae-16ded0d8622e-kube-api-access-7m6w4\") pod \"keystone-80cb-account-create-qcgm2\" (UID: \"8849e0bd-1093-4779-a8ae-16ded0d8622e\") " pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.627862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run-ovn\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.627939 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-log-ovn\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.628033 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48dd72b8-dbca-45bc-b450-eb05338f8fc0-scripts\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.628121 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.648134 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m6w4\" (UniqueName: \"kubernetes.io/projected/8849e0bd-1093-4779-a8ae-16ded0d8622e-kube-api-access-7m6w4\") pod \"keystone-80cb-account-create-qcgm2\" (UID: \"8849e0bd-1093-4779-a8ae-16ded0d8622e\") " pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.699603 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.729429 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jflb\" (UniqueName: \"kubernetes.io/projected/48dd72b8-dbca-45bc-b450-eb05338f8fc0-kube-api-access-9jflb\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.729853 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run-ovn\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.729930 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-log-ovn\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.730048 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48dd72b8-dbca-45bc-b450-eb05338f8fc0-scripts\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.730143 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.730521 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.730575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run-ovn\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.730537 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-log-ovn\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.733002 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48dd72b8-dbca-45bc-b450-eb05338f8fc0-scripts\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.747831 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e50b-account-create-vl6ml"] Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.753207 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.756821 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.767871 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jflb\" (UniqueName: \"kubernetes.io/projected/48dd72b8-dbca-45bc-b450-eb05338f8fc0-kube-api-access-9jflb\") pod \"ovn-controller-gpsfr-config-tj2zs\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.775268 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e50b-account-create-vl6ml"] Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.781600 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:23 crc kubenswrapper[4758]: I1004 11:06:23.931964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-674qw\" (UniqueName: \"kubernetes.io/projected/c3c751e0-cd94-4006-8492-271f1125f266-kube-api-access-674qw\") pod \"placement-e50b-account-create-vl6ml\" (UID: \"c3c751e0-cd94-4006-8492-271f1125f266\") " pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.033770 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-674qw\" (UniqueName: \"kubernetes.io/projected/c3c751e0-cd94-4006-8492-271f1125f266-kube-api-access-674qw\") pod \"placement-e50b-account-create-vl6ml\" (UID: \"c3c751e0-cd94-4006-8492-271f1125f266\") " pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.053516 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-674qw\" (UniqueName: \"kubernetes.io/projected/c3c751e0-cd94-4006-8492-271f1125f266-kube-api-access-674qw\") pod \"placement-e50b-account-create-vl6ml\" (UID: \"c3c751e0-cd94-4006-8492-271f1125f266\") " pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.175728 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.228829 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gpsfr-config-tj2zs"] Oct 04 11:06:24 crc kubenswrapper[4758]: W1004 11:06:24.234636 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48dd72b8_dbca_45bc_b450_eb05338f8fc0.slice/crio-45e328431cb949fa9e226e326e1f01acc0492fd9a00eb95b0a126a9154e778c2 WatchSource:0}: Error finding container 45e328431cb949fa9e226e326e1f01acc0492fd9a00eb95b0a126a9154e778c2: Status 404 returned error can't find the container with id 45e328431cb949fa9e226e326e1f01acc0492fd9a00eb95b0a126a9154e778c2 Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.348345 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-80cb-account-create-qcgm2"] Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.426820 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.543786 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9sqd\" (UniqueName: \"kubernetes.io/projected/a127cd92-1013-49e5-b5ec-832428b0b113-kube-api-access-x9sqd\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.543864 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a127cd92-1013-49e5-b5ec-832428b0b113-etc-swift\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.543926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-ring-data-devices\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.544727 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.546993 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a127cd92-1013-49e5-b5ec-832428b0b113-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.547006 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-dispersionconf\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.547091 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-scripts\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.547150 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-combined-ca-bundle\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.547183 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-swiftconf\") pod \"a127cd92-1013-49e5-b5ec-832428b0b113\" (UID: \"a127cd92-1013-49e5-b5ec-832428b0b113\") " Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.547687 4758 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a127cd92-1013-49e5-b5ec-832428b0b113-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.547814 4758 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.556571 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a127cd92-1013-49e5-b5ec-832428b0b113-kube-api-access-x9sqd" (OuterVolumeSpecName: "kube-api-access-x9sqd") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "kube-api-access-x9sqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.558927 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e50b-account-create-vl6ml"] Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.568345 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.600140 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.601497 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.609165 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-scripts" (OuterVolumeSpecName: "scripts") pod "a127cd92-1013-49e5-b5ec-832428b0b113" (UID: "a127cd92-1013-49e5-b5ec-832428b0b113"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.649052 4758 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.649223 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9sqd\" (UniqueName: \"kubernetes.io/projected/a127cd92-1013-49e5-b5ec-832428b0b113-kube-api-access-x9sqd\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.649343 4758 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.649444 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a127cd92-1013-49e5-b5ec-832428b0b113-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.649546 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a127cd92-1013-49e5-b5ec-832428b0b113-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.953676 4758 generic.go:334] "Generic (PLEG): container finished" podID="48dd72b8-dbca-45bc-b450-eb05338f8fc0" containerID="99c77c2e50490db809ce87a8529b681d0f7a43ec57af42e4eb421689a774e199" exitCode=0 Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.953752 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gpsfr-config-tj2zs" event={"ID":"48dd72b8-dbca-45bc-b450-eb05338f8fc0","Type":"ContainerDied","Data":"99c77c2e50490db809ce87a8529b681d0f7a43ec57af42e4eb421689a774e199"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.953785 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gpsfr-config-tj2zs" event={"ID":"48dd72b8-dbca-45bc-b450-eb05338f8fc0","Type":"ContainerStarted","Data":"45e328431cb949fa9e226e326e1f01acc0492fd9a00eb95b0a126a9154e778c2"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.955949 4758 generic.go:334] "Generic (PLEG): container finished" podID="c3c751e0-cd94-4006-8492-271f1125f266" containerID="07329e5ec7c1aa1806f147ad17c098c9b755affa9b5b56374e764083d2023e85" exitCode=0 Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.956021 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e50b-account-create-vl6ml" event={"ID":"c3c751e0-cd94-4006-8492-271f1125f266","Type":"ContainerDied","Data":"07329e5ec7c1aa1806f147ad17c098c9b755affa9b5b56374e764083d2023e85"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.956059 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e50b-account-create-vl6ml" event={"ID":"c3c751e0-cd94-4006-8492-271f1125f266","Type":"ContainerStarted","Data":"dc844ed492b25087e6383ca5b93c117677db2e1b47253f79cd1888e8efcaf4b3"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.958453 4758 generic.go:334] "Generic (PLEG): container finished" podID="8849e0bd-1093-4779-a8ae-16ded0d8622e" containerID="85fff0c1ee73d1a1c33f5c599c71c593ae99077e09c4e0315b8736dfa9fa4b40" exitCode=0 Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.958508 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-80cb-account-create-qcgm2" event={"ID":"8849e0bd-1093-4779-a8ae-16ded0d8622e","Type":"ContainerDied","Data":"85fff0c1ee73d1a1c33f5c599c71c593ae99077e09c4e0315b8736dfa9fa4b40"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.958525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-80cb-account-create-qcgm2" event={"ID":"8849e0bd-1093-4779-a8ae-16ded0d8622e","Type":"ContainerStarted","Data":"e8f5eb8c23b8fe1e79b4f650cad79dcf2113a4d361aae8cff8d980351bb14ded"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.960582 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9wqh" event={"ID":"a127cd92-1013-49e5-b5ec-832428b0b113","Type":"ContainerDied","Data":"d705309ce1251d4e430875a84ae56431cd5078b14d1a4ad9005626169ee98378"} Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.960609 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d705309ce1251d4e430875a84ae56431cd5078b14d1a4ad9005626169ee98378" Oct 04 11:06:24 crc kubenswrapper[4758]: I1004 11:06:24.960661 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9wqh" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.442718 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.448739 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.454470 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.485917 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run\") pod \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.486017 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m6w4\" (UniqueName: \"kubernetes.io/projected/8849e0bd-1093-4779-a8ae-16ded0d8622e-kube-api-access-7m6w4\") pod \"8849e0bd-1093-4779-a8ae-16ded0d8622e\" (UID: \"8849e0bd-1093-4779-a8ae-16ded0d8622e\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.486137 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run" (OuterVolumeSpecName: "var-run") pod "48dd72b8-dbca-45bc-b450-eb05338f8fc0" (UID: "48dd72b8-dbca-45bc-b450-eb05338f8fc0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.486643 4758 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.492032 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8849e0bd-1093-4779-a8ae-16ded0d8622e-kube-api-access-7m6w4" (OuterVolumeSpecName: "kube-api-access-7m6w4") pod "8849e0bd-1093-4779-a8ae-16ded0d8622e" (UID: "8849e0bd-1093-4779-a8ae-16ded0d8622e"). InnerVolumeSpecName "kube-api-access-7m6w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.587351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48dd72b8-dbca-45bc-b450-eb05338f8fc0-scripts\") pod \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.587424 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jflb\" (UniqueName: \"kubernetes.io/projected/48dd72b8-dbca-45bc-b450-eb05338f8fc0-kube-api-access-9jflb\") pod \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.587457 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run-ovn\") pod \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.587478 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-log-ovn\") pod \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\" (UID: \"48dd72b8-dbca-45bc-b450-eb05338f8fc0\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.587509 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-674qw\" (UniqueName: \"kubernetes.io/projected/c3c751e0-cd94-4006-8492-271f1125f266-kube-api-access-674qw\") pod \"c3c751e0-cd94-4006-8492-271f1125f266\" (UID: \"c3c751e0-cd94-4006-8492-271f1125f266\") " Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.587819 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m6w4\" (UniqueName: \"kubernetes.io/projected/8849e0bd-1093-4779-a8ae-16ded0d8622e-kube-api-access-7m6w4\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.588048 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "48dd72b8-dbca-45bc-b450-eb05338f8fc0" (UID: "48dd72b8-dbca-45bc-b450-eb05338f8fc0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.588074 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "48dd72b8-dbca-45bc-b450-eb05338f8fc0" (UID: "48dd72b8-dbca-45bc-b450-eb05338f8fc0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.588567 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48dd72b8-dbca-45bc-b450-eb05338f8fc0-scripts" (OuterVolumeSpecName: "scripts") pod "48dd72b8-dbca-45bc-b450-eb05338f8fc0" (UID: "48dd72b8-dbca-45bc-b450-eb05338f8fc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.591547 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48dd72b8-dbca-45bc-b450-eb05338f8fc0-kube-api-access-9jflb" (OuterVolumeSpecName: "kube-api-access-9jflb") pod "48dd72b8-dbca-45bc-b450-eb05338f8fc0" (UID: "48dd72b8-dbca-45bc-b450-eb05338f8fc0"). InnerVolumeSpecName "kube-api-access-9jflb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.593924 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c751e0-cd94-4006-8492-271f1125f266-kube-api-access-674qw" (OuterVolumeSpecName: "kube-api-access-674qw") pod "c3c751e0-cd94-4006-8492-271f1125f266" (UID: "c3c751e0-cd94-4006-8492-271f1125f266"). InnerVolumeSpecName "kube-api-access-674qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.688892 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48dd72b8-dbca-45bc-b450-eb05338f8fc0-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.688926 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jflb\" (UniqueName: \"kubernetes.io/projected/48dd72b8-dbca-45bc-b450-eb05338f8fc0-kube-api-access-9jflb\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.688937 4758 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.688946 4758 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/48dd72b8-dbca-45bc-b450-eb05338f8fc0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.688956 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-674qw\" (UniqueName: \"kubernetes.io/projected/c3c751e0-cd94-4006-8492-271f1125f266-kube-api-access-674qw\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.982912 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gpsfr-config-tj2zs" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.982914 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gpsfr-config-tj2zs" event={"ID":"48dd72b8-dbca-45bc-b450-eb05338f8fc0","Type":"ContainerDied","Data":"45e328431cb949fa9e226e326e1f01acc0492fd9a00eb95b0a126a9154e778c2"} Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.982959 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45e328431cb949fa9e226e326e1f01acc0492fd9a00eb95b0a126a9154e778c2" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.985662 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e50b-account-create-vl6ml" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.985657 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e50b-account-create-vl6ml" event={"ID":"c3c751e0-cd94-4006-8492-271f1125f266","Type":"ContainerDied","Data":"dc844ed492b25087e6383ca5b93c117677db2e1b47253f79cd1888e8efcaf4b3"} Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.985788 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc844ed492b25087e6383ca5b93c117677db2e1b47253f79cd1888e8efcaf4b3" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.986921 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-80cb-account-create-qcgm2" event={"ID":"8849e0bd-1093-4779-a8ae-16ded0d8622e","Type":"ContainerDied","Data":"e8f5eb8c23b8fe1e79b4f650cad79dcf2113a4d361aae8cff8d980351bb14ded"} Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.986951 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8f5eb8c23b8fe1e79b4f650cad79dcf2113a4d361aae8cff8d980351bb14ded" Oct 04 11:06:26 crc kubenswrapper[4758]: I1004 11:06:26.986993 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-80cb-account-create-qcgm2" Oct 04 11:06:27 crc kubenswrapper[4758]: I1004 11:06:27.550689 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gpsfr-config-tj2zs"] Oct 04 11:06:27 crc kubenswrapper[4758]: I1004 11:06:27.557374 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gpsfr-config-tj2zs"] Oct 04 11:06:27 crc kubenswrapper[4758]: I1004 11:06:27.996650 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gpsfr" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987187 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-76e3-account-create-xn8n5"] Oct 04 11:06:28 crc kubenswrapper[4758]: E1004 11:06:28.987493 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a127cd92-1013-49e5-b5ec-832428b0b113" containerName="swift-ring-rebalance" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987508 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a127cd92-1013-49e5-b5ec-832428b0b113" containerName="swift-ring-rebalance" Oct 04 11:06:28 crc kubenswrapper[4758]: E1004 11:06:28.987526 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c751e0-cd94-4006-8492-271f1125f266" containerName="mariadb-account-create" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987533 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c751e0-cd94-4006-8492-271f1125f266" containerName="mariadb-account-create" Oct 04 11:06:28 crc kubenswrapper[4758]: E1004 11:06:28.987552 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8849e0bd-1093-4779-a8ae-16ded0d8622e" containerName="mariadb-account-create" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987560 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8849e0bd-1093-4779-a8ae-16ded0d8622e" containerName="mariadb-account-create" Oct 04 11:06:28 crc kubenswrapper[4758]: E1004 11:06:28.987572 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48dd72b8-dbca-45bc-b450-eb05338f8fc0" containerName="ovn-config" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987578 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="48dd72b8-dbca-45bc-b450-eb05338f8fc0" containerName="ovn-config" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987736 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="48dd72b8-dbca-45bc-b450-eb05338f8fc0" containerName="ovn-config" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987750 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a127cd92-1013-49e5-b5ec-832428b0b113" containerName="swift-ring-rebalance" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987762 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c751e0-cd94-4006-8492-271f1125f266" containerName="mariadb-account-create" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.987772 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8849e0bd-1093-4779-a8ae-16ded0d8622e" containerName="mariadb-account-create" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.988346 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.990182 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 11:06:28 crc kubenswrapper[4758]: I1004 11:06:28.999865 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-76e3-account-create-xn8n5"] Oct 04 11:06:29 crc kubenswrapper[4758]: I1004 11:06:29.029308 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj7rk\" (UniqueName: \"kubernetes.io/projected/48a3538c-a18f-41b2-87f6-59240be0d9f8-kube-api-access-bj7rk\") pod \"glance-76e3-account-create-xn8n5\" (UID: \"48a3538c-a18f-41b2-87f6-59240be0d9f8\") " pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:29 crc kubenswrapper[4758]: I1004 11:06:29.130796 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj7rk\" (UniqueName: \"kubernetes.io/projected/48a3538c-a18f-41b2-87f6-59240be0d9f8-kube-api-access-bj7rk\") pod \"glance-76e3-account-create-xn8n5\" (UID: \"48a3538c-a18f-41b2-87f6-59240be0d9f8\") " pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:29 crc kubenswrapper[4758]: I1004 11:06:29.154635 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj7rk\" (UniqueName: \"kubernetes.io/projected/48a3538c-a18f-41b2-87f6-59240be0d9f8-kube-api-access-bj7rk\") pod \"glance-76e3-account-create-xn8n5\" (UID: \"48a3538c-a18f-41b2-87f6-59240be0d9f8\") " pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:29 crc kubenswrapper[4758]: I1004 11:06:29.310335 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:29 crc kubenswrapper[4758]: I1004 11:06:29.335761 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48dd72b8-dbca-45bc-b450-eb05338f8fc0" path="/var/lib/kubelet/pods/48dd72b8-dbca-45bc-b450-eb05338f8fc0/volumes" Oct 04 11:06:29 crc kubenswrapper[4758]: I1004 11:06:29.756112 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-76e3-account-create-xn8n5"] Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.011400 4758 generic.go:334] "Generic (PLEG): container finished" podID="48a3538c-a18f-41b2-87f6-59240be0d9f8" containerID="6023cda852b544c11a75f84fcbe94b1fc1752d7b23f990722a98f3224c951939" exitCode=0 Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.011580 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-76e3-account-create-xn8n5" event={"ID":"48a3538c-a18f-41b2-87f6-59240be0d9f8","Type":"ContainerDied","Data":"6023cda852b544c11a75f84fcbe94b1fc1752d7b23f990722a98f3224c951939"} Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.011769 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-76e3-account-create-xn8n5" event={"ID":"48a3538c-a18f-41b2-87f6-59240be0d9f8","Type":"ContainerStarted","Data":"01a7c27dd2c908a4be30c1cd0e5bb4f68482fe024c0a14cbbbb38dcc19fb844b"} Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.230367 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.510291 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.713181 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-t5kz9"] Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.714117 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.781701 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t5kz9"] Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.824210 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8s4md"] Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.825646 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.843710 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8s4md"] Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.860774 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r74d7\" (UniqueName: \"kubernetes.io/projected/246cc3c0-2cb3-4def-bac8-54aa56554813-kube-api-access-r74d7\") pod \"cinder-db-create-t5kz9\" (UID: \"246cc3c0-2cb3-4def-bac8-54aa56554813\") " pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.961840 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r74d7\" (UniqueName: \"kubernetes.io/projected/246cc3c0-2cb3-4def-bac8-54aa56554813-kube-api-access-r74d7\") pod \"cinder-db-create-t5kz9\" (UID: \"246cc3c0-2cb3-4def-bac8-54aa56554813\") " pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.962067 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgq25\" (UniqueName: \"kubernetes.io/projected/c40ad036-101f-4e20-b049-cd53739e957f-kube-api-access-sgq25\") pod \"barbican-db-create-8s4md\" (UID: \"c40ad036-101f-4e20-b049-cd53739e957f\") " pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:30 crc kubenswrapper[4758]: I1004 11:06:30.982661 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r74d7\" (UniqueName: \"kubernetes.io/projected/246cc3c0-2cb3-4def-bac8-54aa56554813-kube-api-access-r74d7\") pod \"cinder-db-create-t5kz9\" (UID: \"246cc3c0-2cb3-4def-bac8-54aa56554813\") " pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.024457 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-bgthn"] Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.030284 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.035092 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.041245 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bgthn"] Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.064467 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgq25\" (UniqueName: \"kubernetes.io/projected/c40ad036-101f-4e20-b049-cd53739e957f-kube-api-access-sgq25\") pod \"barbican-db-create-8s4md\" (UID: \"c40ad036-101f-4e20-b049-cd53739e957f\") " pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.103181 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgq25\" (UniqueName: \"kubernetes.io/projected/c40ad036-101f-4e20-b049-cd53739e957f-kube-api-access-sgq25\") pod \"barbican-db-create-8s4md\" (UID: \"c40ad036-101f-4e20-b049-cd53739e957f\") " pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.145637 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.166393 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfxsg\" (UniqueName: \"kubernetes.io/projected/c573ca52-55dd-4efd-9227-35d71b1f5268-kube-api-access-gfxsg\") pod \"neutron-db-create-bgthn\" (UID: \"c573ca52-55dd-4efd-9227-35d71b1f5268\") " pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.249248 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.249289 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.268319 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfxsg\" (UniqueName: \"kubernetes.io/projected/c573ca52-55dd-4efd-9227-35d71b1f5268-kube-api-access-gfxsg\") pod \"neutron-db-create-bgthn\" (UID: \"c573ca52-55dd-4efd-9227-35d71b1f5268\") " pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.286939 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfxsg\" (UniqueName: \"kubernetes.io/projected/c573ca52-55dd-4efd-9227-35d71b1f5268-kube-api-access-gfxsg\") pod \"neutron-db-create-bgthn\" (UID: \"c573ca52-55dd-4efd-9227-35d71b1f5268\") " pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.397869 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.411773 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t5kz9"] Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.465426 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.574492 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj7rk\" (UniqueName: \"kubernetes.io/projected/48a3538c-a18f-41b2-87f6-59240be0d9f8-kube-api-access-bj7rk\") pod \"48a3538c-a18f-41b2-87f6-59240be0d9f8\" (UID: \"48a3538c-a18f-41b2-87f6-59240be0d9f8\") " Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.585318 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48a3538c-a18f-41b2-87f6-59240be0d9f8-kube-api-access-bj7rk" (OuterVolumeSpecName: "kube-api-access-bj7rk") pod "48a3538c-a18f-41b2-87f6-59240be0d9f8" (UID: "48a3538c-a18f-41b2-87f6-59240be0d9f8"). InnerVolumeSpecName "kube-api-access-bj7rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.677409 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj7rk\" (UniqueName: \"kubernetes.io/projected/48a3538c-a18f-41b2-87f6-59240be0d9f8-kube-api-access-bj7rk\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:31 crc kubenswrapper[4758]: I1004 11:06:31.747472 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8s4md"] Oct 04 11:06:31 crc kubenswrapper[4758]: W1004 11:06:31.748303 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc40ad036_101f_4e20_b049_cd53739e957f.slice/crio-9bad937f418c8b01ae36181e4a80b390acfba7de8ec4a4087085cb1ce6c09828 WatchSource:0}: Error finding container 9bad937f418c8b01ae36181e4a80b390acfba7de8ec4a4087085cb1ce6c09828: Status 404 returned error can't find the container with id 9bad937f418c8b01ae36181e4a80b390acfba7de8ec4a4087085cb1ce6c09828 Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.028234 4758 generic.go:334] "Generic (PLEG): container finished" podID="246cc3c0-2cb3-4def-bac8-54aa56554813" containerID="7e0830d3b36673987fc3a4abd269a946c10ac75eabb8e0f19fba2aca12f5cae8" exitCode=0 Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.028333 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t5kz9" event={"ID":"246cc3c0-2cb3-4def-bac8-54aa56554813","Type":"ContainerDied","Data":"7e0830d3b36673987fc3a4abd269a946c10ac75eabb8e0f19fba2aca12f5cae8"} Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.028382 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t5kz9" event={"ID":"246cc3c0-2cb3-4def-bac8-54aa56554813","Type":"ContainerStarted","Data":"ca63dd05e5c3dc657e166ac22fa5fc50f9074d2a11b94197ce5caa01d4cbd5fa"} Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.030086 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-76e3-account-create-xn8n5" event={"ID":"48a3538c-a18f-41b2-87f6-59240be0d9f8","Type":"ContainerDied","Data":"01a7c27dd2c908a4be30c1cd0e5bb4f68482fe024c0a14cbbbb38dcc19fb844b"} Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.030123 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-76e3-account-create-xn8n5" Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.030131 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01a7c27dd2c908a4be30c1cd0e5bb4f68482fe024c0a14cbbbb38dcc19fb844b" Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.031649 4758 generic.go:334] "Generic (PLEG): container finished" podID="c40ad036-101f-4e20-b049-cd53739e957f" containerID="cf3bbd33dc6d0dff561e88c7fbae037cd665c7080d779b7c10bae31d0b09175b" exitCode=0 Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.031683 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8s4md" event={"ID":"c40ad036-101f-4e20-b049-cd53739e957f","Type":"ContainerDied","Data":"cf3bbd33dc6d0dff561e88c7fbae037cd665c7080d779b7c10bae31d0b09175b"} Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.031702 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8s4md" event={"ID":"c40ad036-101f-4e20-b049-cd53739e957f","Type":"ContainerStarted","Data":"9bad937f418c8b01ae36181e4a80b390acfba7de8ec4a4087085cb1ce6c09828"} Oct 04 11:06:32 crc kubenswrapper[4758]: I1004 11:06:32.076234 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bgthn"] Oct 04 11:06:32 crc kubenswrapper[4758]: W1004 11:06:32.117760 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc573ca52_55dd_4efd_9227_35d71b1f5268.slice/crio-c9638aafc2758af681cea44f5e578304f28210f71af7eec454b2aacfa5dfb0b0 WatchSource:0}: Error finding container c9638aafc2758af681cea44f5e578304f28210f71af7eec454b2aacfa5dfb0b0: Status 404 returned error can't find the container with id c9638aafc2758af681cea44f5e578304f28210f71af7eec454b2aacfa5dfb0b0 Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.054659 4758 generic.go:334] "Generic (PLEG): container finished" podID="c573ca52-55dd-4efd-9227-35d71b1f5268" containerID="d2bb6e8e87efc9456c81cd214b0a3f2b4f2f2aba2969b1474ca5b45b78bffd85" exitCode=0 Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.055869 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bgthn" event={"ID":"c573ca52-55dd-4efd-9227-35d71b1f5268","Type":"ContainerDied","Data":"d2bb6e8e87efc9456c81cd214b0a3f2b4f2f2aba2969b1474ca5b45b78bffd85"} Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.056008 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bgthn" event={"ID":"c573ca52-55dd-4efd-9227-35d71b1f5268","Type":"ContainerStarted","Data":"c9638aafc2758af681cea44f5e578304f28210f71af7eec454b2aacfa5dfb0b0"} Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.520714 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.525440 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.607598 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-crlgt"] Oct 04 11:06:33 crc kubenswrapper[4758]: E1004 11:06:33.607882 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40ad036-101f-4e20-b049-cd53739e957f" containerName="mariadb-database-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.607893 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40ad036-101f-4e20-b049-cd53739e957f" containerName="mariadb-database-create" Oct 04 11:06:33 crc kubenswrapper[4758]: E1004 11:06:33.607907 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a3538c-a18f-41b2-87f6-59240be0d9f8" containerName="mariadb-account-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.607913 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a3538c-a18f-41b2-87f6-59240be0d9f8" containerName="mariadb-account-create" Oct 04 11:06:33 crc kubenswrapper[4758]: E1004 11:06:33.607932 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="246cc3c0-2cb3-4def-bac8-54aa56554813" containerName="mariadb-database-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.607937 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="246cc3c0-2cb3-4def-bac8-54aa56554813" containerName="mariadb-database-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.608074 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="246cc3c0-2cb3-4def-bac8-54aa56554813" containerName="mariadb-database-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.608756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r74d7\" (UniqueName: \"kubernetes.io/projected/246cc3c0-2cb3-4def-bac8-54aa56554813-kube-api-access-r74d7\") pod \"246cc3c0-2cb3-4def-bac8-54aa56554813\" (UID: \"246cc3c0-2cb3-4def-bac8-54aa56554813\") " Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.609419 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40ad036-101f-4e20-b049-cd53739e957f" containerName="mariadb-database-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.609433 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a3538c-a18f-41b2-87f6-59240be0d9f8" containerName="mariadb-account-create" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.609959 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.617450 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/246cc3c0-2cb3-4def-bac8-54aa56554813-kube-api-access-r74d7" (OuterVolumeSpecName: "kube-api-access-r74d7") pod "246cc3c0-2cb3-4def-bac8-54aa56554813" (UID: "246cc3c0-2cb3-4def-bac8-54aa56554813"). InnerVolumeSpecName "kube-api-access-r74d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.618037 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.618225 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4n6pt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.618330 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.618435 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.622968 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-crlgt"] Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.711092 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgq25\" (UniqueName: \"kubernetes.io/projected/c40ad036-101f-4e20-b049-cd53739e957f-kube-api-access-sgq25\") pod \"c40ad036-101f-4e20-b049-cd53739e957f\" (UID: \"c40ad036-101f-4e20-b049-cd53739e957f\") " Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.711758 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-config-data\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.711836 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-combined-ca-bundle\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.711901 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgctw\" (UniqueName: \"kubernetes.io/projected/157f102f-c0ec-4e03-97cb-9b45948c82ea-kube-api-access-qgctw\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.711947 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r74d7\" (UniqueName: \"kubernetes.io/projected/246cc3c0-2cb3-4def-bac8-54aa56554813-kube-api-access-r74d7\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.714651 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40ad036-101f-4e20-b049-cd53739e957f-kube-api-access-sgq25" (OuterVolumeSpecName: "kube-api-access-sgq25") pod "c40ad036-101f-4e20-b049-cd53739e957f" (UID: "c40ad036-101f-4e20-b049-cd53739e957f"). InnerVolumeSpecName "kube-api-access-sgq25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.812702 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-combined-ca-bundle\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.812775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgctw\" (UniqueName: \"kubernetes.io/projected/157f102f-c0ec-4e03-97cb-9b45948c82ea-kube-api-access-qgctw\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.812830 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-config-data\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.812982 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgq25\" (UniqueName: \"kubernetes.io/projected/c40ad036-101f-4e20-b049-cd53739e957f-kube-api-access-sgq25\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.818091 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-config-data\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.829380 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-combined-ca-bundle\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.833049 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgctw\" (UniqueName: \"kubernetes.io/projected/157f102f-c0ec-4e03-97cb-9b45948c82ea-kube-api-access-qgctw\") pod \"keystone-db-sync-crlgt\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:33 crc kubenswrapper[4758]: I1004 11:06:33.957781 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-crlgt" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.074742 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8s4md" event={"ID":"c40ad036-101f-4e20-b049-cd53739e957f","Type":"ContainerDied","Data":"9bad937f418c8b01ae36181e4a80b390acfba7de8ec4a4087085cb1ce6c09828"} Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.074777 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bad937f418c8b01ae36181e4a80b390acfba7de8ec4a4087085cb1ce6c09828" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.074840 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8s4md" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.077078 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t5kz9" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.081147 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t5kz9" event={"ID":"246cc3c0-2cb3-4def-bac8-54aa56554813","Type":"ContainerDied","Data":"ca63dd05e5c3dc657e166ac22fa5fc50f9074d2a11b94197ce5caa01d4cbd5fa"} Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.081179 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca63dd05e5c3dc657e166ac22fa5fc50f9074d2a11b94197ce5caa01d4cbd5fa" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.114285 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-trp6p"] Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.117411 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.122811 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.122999 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8w5vq" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.135159 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-trp6p"] Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.218213 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-combined-ca-bundle\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.218260 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82bdl\" (UniqueName: \"kubernetes.io/projected/bfe10db6-4338-48ab-b744-18213468c240-kube-api-access-82bdl\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.218290 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-db-sync-config-data\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.218315 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-config-data\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.321185 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-config-data\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.322516 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-combined-ca-bundle\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.322579 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82bdl\" (UniqueName: \"kubernetes.io/projected/bfe10db6-4338-48ab-b744-18213468c240-kube-api-access-82bdl\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.322616 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-db-sync-config-data\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.330400 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-db-sync-config-data\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.330586 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-combined-ca-bundle\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.333725 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-config-data\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.346076 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82bdl\" (UniqueName: \"kubernetes.io/projected/bfe10db6-4338-48ab-b744-18213468c240-kube-api-access-82bdl\") pod \"glance-db-sync-trp6p\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.364686 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.422128 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-crlgt"] Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.454872 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-trp6p" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.525873 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfxsg\" (UniqueName: \"kubernetes.io/projected/c573ca52-55dd-4efd-9227-35d71b1f5268-kube-api-access-gfxsg\") pod \"c573ca52-55dd-4efd-9227-35d71b1f5268\" (UID: \"c573ca52-55dd-4efd-9227-35d71b1f5268\") " Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.530422 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c573ca52-55dd-4efd-9227-35d71b1f5268-kube-api-access-gfxsg" (OuterVolumeSpecName: "kube-api-access-gfxsg") pod "c573ca52-55dd-4efd-9227-35d71b1f5268" (UID: "c573ca52-55dd-4efd-9227-35d71b1f5268"). InnerVolumeSpecName "kube-api-access-gfxsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:34 crc kubenswrapper[4758]: I1004 11:06:34.627585 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfxsg\" (UniqueName: \"kubernetes.io/projected/c573ca52-55dd-4efd-9227-35d71b1f5268-kube-api-access-gfxsg\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:35 crc kubenswrapper[4758]: I1004 11:06:35.012364 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-trp6p"] Oct 04 11:06:35 crc kubenswrapper[4758]: W1004 11:06:35.025233 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfe10db6_4338_48ab_b744_18213468c240.slice/crio-028432c4df9435b2296627b31431737cd92de31914ae33e1620b16f5d821ea87 WatchSource:0}: Error finding container 028432c4df9435b2296627b31431737cd92de31914ae33e1620b16f5d821ea87: Status 404 returned error can't find the container with id 028432c4df9435b2296627b31431737cd92de31914ae33e1620b16f5d821ea87 Oct 04 11:06:35 crc kubenswrapper[4758]: I1004 11:06:35.085408 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-crlgt" event={"ID":"157f102f-c0ec-4e03-97cb-9b45948c82ea","Type":"ContainerStarted","Data":"93944d491d101fc3a574254df8f614dff30703fdc4720813814bf693379bb577"} Oct 04 11:06:35 crc kubenswrapper[4758]: I1004 11:06:35.086671 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bgthn" event={"ID":"c573ca52-55dd-4efd-9227-35d71b1f5268","Type":"ContainerDied","Data":"c9638aafc2758af681cea44f5e578304f28210f71af7eec454b2aacfa5dfb0b0"} Oct 04 11:06:35 crc kubenswrapper[4758]: I1004 11:06:35.086727 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9638aafc2758af681cea44f5e578304f28210f71af7eec454b2aacfa5dfb0b0" Oct 04 11:06:35 crc kubenswrapper[4758]: I1004 11:06:35.086935 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bgthn" Oct 04 11:06:35 crc kubenswrapper[4758]: I1004 11:06:35.087456 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-trp6p" event={"ID":"bfe10db6-4338-48ab-b744-18213468c240","Type":"ContainerStarted","Data":"028432c4df9435b2296627b31431737cd92de31914ae33e1620b16f5d821ea87"} Oct 04 11:06:37 crc kubenswrapper[4758]: I1004 11:06:37.779606 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:37 crc kubenswrapper[4758]: I1004 11:06:37.786575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3ed3476-2568-4c27-ac70-87d732389beb-etc-swift\") pod \"swift-storage-0\" (UID: \"d3ed3476-2568-4c27-ac70-87d732389beb\") " pod="openstack/swift-storage-0" Oct 04 11:06:37 crc kubenswrapper[4758]: I1004 11:06:37.896911 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.682130 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-bbb3-account-create-2frqn"] Oct 04 11:06:40 crc kubenswrapper[4758]: E1004 11:06:40.682983 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c573ca52-55dd-4efd-9227-35d71b1f5268" containerName="mariadb-database-create" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.682998 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c573ca52-55dd-4efd-9227-35d71b1f5268" containerName="mariadb-database-create" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.683264 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c573ca52-55dd-4efd-9227-35d71b1f5268" containerName="mariadb-database-create" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.683823 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.685935 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.707186 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bbb3-account-create-2frqn"] Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.807187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fp4q\" (UniqueName: \"kubernetes.io/projected/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4-kube-api-access-5fp4q\") pod \"cinder-bbb3-account-create-2frqn\" (UID: \"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4\") " pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.860373 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6ad0-account-create-6rh8j"] Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.862015 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.864165 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.883864 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6ad0-account-create-6rh8j"] Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.909254 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4qz4\" (UniqueName: \"kubernetes.io/projected/0a3ed6de-8959-441c-bf24-025f5e7c185d-kube-api-access-l4qz4\") pod \"barbican-6ad0-account-create-6rh8j\" (UID: \"0a3ed6de-8959-441c-bf24-025f5e7c185d\") " pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.909560 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fp4q\" (UniqueName: \"kubernetes.io/projected/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4-kube-api-access-5fp4q\") pod \"cinder-bbb3-account-create-2frqn\" (UID: \"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4\") " pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:40 crc kubenswrapper[4758]: I1004 11:06:40.928017 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fp4q\" (UniqueName: \"kubernetes.io/projected/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4-kube-api-access-5fp4q\") pod \"cinder-bbb3-account-create-2frqn\" (UID: \"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4\") " pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.010725 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4qz4\" (UniqueName: \"kubernetes.io/projected/0a3ed6de-8959-441c-bf24-025f5e7c185d-kube-api-access-l4qz4\") pod \"barbican-6ad0-account-create-6rh8j\" (UID: \"0a3ed6de-8959-441c-bf24-025f5e7c185d\") " pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.010926 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.026387 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4qz4\" (UniqueName: \"kubernetes.io/projected/0a3ed6de-8959-441c-bf24-025f5e7c185d-kube-api-access-l4qz4\") pod \"barbican-6ad0-account-create-6rh8j\" (UID: \"0a3ed6de-8959-441c-bf24-025f5e7c185d\") " pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.057197 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-4285-account-create-rzjvs"] Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.058373 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.060186 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.072610 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-4285-account-create-rzjvs"] Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.112471 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c45sv\" (UniqueName: \"kubernetes.io/projected/ab228585-a058-4d2a-aac4-6dd6bbf30852-kube-api-access-c45sv\") pod \"neutron-4285-account-create-rzjvs\" (UID: \"ab228585-a058-4d2a-aac4-6dd6bbf30852\") " pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.180749 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.213329 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c45sv\" (UniqueName: \"kubernetes.io/projected/ab228585-a058-4d2a-aac4-6dd6bbf30852-kube-api-access-c45sv\") pod \"neutron-4285-account-create-rzjvs\" (UID: \"ab228585-a058-4d2a-aac4-6dd6bbf30852\") " pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.236447 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c45sv\" (UniqueName: \"kubernetes.io/projected/ab228585-a058-4d2a-aac4-6dd6bbf30852-kube-api-access-c45sv\") pod \"neutron-4285-account-create-rzjvs\" (UID: \"ab228585-a058-4d2a-aac4-6dd6bbf30852\") " pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:41 crc kubenswrapper[4758]: I1004 11:06:41.386361 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:52 crc kubenswrapper[4758]: E1004 11:06:52.784216 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Oct 04 11:06:52 crc kubenswrapper[4758]: E1004 11:06:52.784877 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qgctw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-crlgt_openstack(157f102f-c0ec-4e03-97cb-9b45948c82ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:06:52 crc kubenswrapper[4758]: E1004 11:06:52.786829 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-crlgt" podUID="157f102f-c0ec-4e03-97cb-9b45948c82ea" Oct 04 11:06:53 crc kubenswrapper[4758]: E1004 11:06:53.247186 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-crlgt" podUID="157f102f-c0ec-4e03-97cb-9b45948c82ea" Oct 04 11:06:53 crc kubenswrapper[4758]: E1004 11:06:53.854572 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Oct 04 11:06:53 crc kubenswrapper[4758]: E1004 11:06:53.854763 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-82bdl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-trp6p_openstack(bfe10db6-4338-48ab-b744-18213468c240): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:06:53 crc kubenswrapper[4758]: E1004 11:06:53.856334 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-trp6p" podUID="bfe10db6-4338-48ab-b744-18213468c240" Oct 04 11:06:54 crc kubenswrapper[4758]: I1004 11:06:54.146236 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-4285-account-create-rzjvs"] Oct 04 11:06:54 crc kubenswrapper[4758]: I1004 11:06:54.254691 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4285-account-create-rzjvs" event={"ID":"ab228585-a058-4d2a-aac4-6dd6bbf30852","Type":"ContainerStarted","Data":"3377fb05556c966d97f90e5347b8f03c8eea9259bbe72a38f005cdcebacfbbb1"} Oct 04 11:06:54 crc kubenswrapper[4758]: E1004 11:06:54.256506 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-trp6p" podUID="bfe10db6-4338-48ab-b744-18213468c240" Oct 04 11:06:54 crc kubenswrapper[4758]: I1004 11:06:54.402393 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6ad0-account-create-6rh8j"] Oct 04 11:06:54 crc kubenswrapper[4758]: I1004 11:06:54.409604 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-bbb3-account-create-2frqn"] Oct 04 11:06:54 crc kubenswrapper[4758]: W1004 11:06:54.417190 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fa5cd7e_8df4_42ad_bf1d_10728e3fb2c4.slice/crio-1f008d96ef7e7319f7129146f5248a1dc376505e9d39ff950096d0322d5dbd08 WatchSource:0}: Error finding container 1f008d96ef7e7319f7129146f5248a1dc376505e9d39ff950096d0322d5dbd08: Status 404 returned error can't find the container with id 1f008d96ef7e7319f7129146f5248a1dc376505e9d39ff950096d0322d5dbd08 Oct 04 11:06:54 crc kubenswrapper[4758]: I1004 11:06:54.583171 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 11:06:54 crc kubenswrapper[4758]: W1004 11:06:54.592018 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3ed3476_2568_4c27_ac70_87d732389beb.slice/crio-f7353dbfde8f287a387fa627a0ff3ec502f2ca9f60a1d1c75e325f3419d5fed2 WatchSource:0}: Error finding container f7353dbfde8f287a387fa627a0ff3ec502f2ca9f60a1d1c75e325f3419d5fed2: Status 404 returned error can't find the container with id f7353dbfde8f287a387fa627a0ff3ec502f2ca9f60a1d1c75e325f3419d5fed2 Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.267299 4758 generic.go:334] "Generic (PLEG): container finished" podID="7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4" containerID="5d5c2ec209501eacb40bda392aafacf21398d2ec28df229e6b6b39fa3fd21693" exitCode=0 Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.267390 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb3-account-create-2frqn" event={"ID":"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4","Type":"ContainerDied","Data":"5d5c2ec209501eacb40bda392aafacf21398d2ec28df229e6b6b39fa3fd21693"} Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.267723 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb3-account-create-2frqn" event={"ID":"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4","Type":"ContainerStarted","Data":"1f008d96ef7e7319f7129146f5248a1dc376505e9d39ff950096d0322d5dbd08"} Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.271327 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"f7353dbfde8f287a387fa627a0ff3ec502f2ca9f60a1d1c75e325f3419d5fed2"} Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.277599 4758 generic.go:334] "Generic (PLEG): container finished" podID="ab228585-a058-4d2a-aac4-6dd6bbf30852" containerID="3d9139b91b63076052894750d8ed65555d0b934b83aeecf48a84686a4b61d02e" exitCode=0 Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.277713 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4285-account-create-rzjvs" event={"ID":"ab228585-a058-4d2a-aac4-6dd6bbf30852","Type":"ContainerDied","Data":"3d9139b91b63076052894750d8ed65555d0b934b83aeecf48a84686a4b61d02e"} Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.281627 4758 generic.go:334] "Generic (PLEG): container finished" podID="0a3ed6de-8959-441c-bf24-025f5e7c185d" containerID="5bc7ddabb624c50b1415f39497cc2b97cebbd2598f9bd3e15deb94dc9ff29879" exitCode=0 Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.281669 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad0-account-create-6rh8j" event={"ID":"0a3ed6de-8959-441c-bf24-025f5e7c185d","Type":"ContainerDied","Data":"5bc7ddabb624c50b1415f39497cc2b97cebbd2598f9bd3e15deb94dc9ff29879"} Oct 04 11:06:55 crc kubenswrapper[4758]: I1004 11:06:55.281694 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad0-account-create-6rh8j" event={"ID":"0a3ed6de-8959-441c-bf24-025f5e7c185d","Type":"ContainerStarted","Data":"24551f8f4814d6826b0303eadf876af02f2ffe76a576dfc043c886127b46a8ce"} Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.730039 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.737981 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.746974 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.902827 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fp4q\" (UniqueName: \"kubernetes.io/projected/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4-kube-api-access-5fp4q\") pod \"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4\" (UID: \"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4\") " Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.903138 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4qz4\" (UniqueName: \"kubernetes.io/projected/0a3ed6de-8959-441c-bf24-025f5e7c185d-kube-api-access-l4qz4\") pod \"0a3ed6de-8959-441c-bf24-025f5e7c185d\" (UID: \"0a3ed6de-8959-441c-bf24-025f5e7c185d\") " Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.903164 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c45sv\" (UniqueName: \"kubernetes.io/projected/ab228585-a058-4d2a-aac4-6dd6bbf30852-kube-api-access-c45sv\") pod \"ab228585-a058-4d2a-aac4-6dd6bbf30852\" (UID: \"ab228585-a058-4d2a-aac4-6dd6bbf30852\") " Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.923325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a3ed6de-8959-441c-bf24-025f5e7c185d-kube-api-access-l4qz4" (OuterVolumeSpecName: "kube-api-access-l4qz4") pod "0a3ed6de-8959-441c-bf24-025f5e7c185d" (UID: "0a3ed6de-8959-441c-bf24-025f5e7c185d"). InnerVolumeSpecName "kube-api-access-l4qz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.923389 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4-kube-api-access-5fp4q" (OuterVolumeSpecName: "kube-api-access-5fp4q") pod "7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4" (UID: "7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4"). InnerVolumeSpecName "kube-api-access-5fp4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:56 crc kubenswrapper[4758]: I1004 11:06:56.928208 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab228585-a058-4d2a-aac4-6dd6bbf30852-kube-api-access-c45sv" (OuterVolumeSpecName: "kube-api-access-c45sv") pod "ab228585-a058-4d2a-aac4-6dd6bbf30852" (UID: "ab228585-a058-4d2a-aac4-6dd6bbf30852"). InnerVolumeSpecName "kube-api-access-c45sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.005559 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fp4q\" (UniqueName: \"kubernetes.io/projected/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4-kube-api-access-5fp4q\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.005588 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4qz4\" (UniqueName: \"kubernetes.io/projected/0a3ed6de-8959-441c-bf24-025f5e7c185d-kube-api-access-l4qz4\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.005598 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c45sv\" (UniqueName: \"kubernetes.io/projected/ab228585-a058-4d2a-aac4-6dd6bbf30852-kube-api-access-c45sv\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.299329 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-4285-account-create-rzjvs" event={"ID":"ab228585-a058-4d2a-aac4-6dd6bbf30852","Type":"ContainerDied","Data":"3377fb05556c966d97f90e5347b8f03c8eea9259bbe72a38f005cdcebacfbbb1"} Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.299379 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3377fb05556c966d97f90e5347b8f03c8eea9259bbe72a38f005cdcebacfbbb1" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.299448 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-4285-account-create-rzjvs" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.302016 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6ad0-account-create-6rh8j" event={"ID":"0a3ed6de-8959-441c-bf24-025f5e7c185d","Type":"ContainerDied","Data":"24551f8f4814d6826b0303eadf876af02f2ffe76a576dfc043c886127b46a8ce"} Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.302074 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24551f8f4814d6826b0303eadf876af02f2ffe76a576dfc043c886127b46a8ce" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.302229 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6ad0-account-create-6rh8j" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.307248 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-bbb3-account-create-2frqn" event={"ID":"7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4","Type":"ContainerDied","Data":"1f008d96ef7e7319f7129146f5248a1dc376505e9d39ff950096d0322d5dbd08"} Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.307288 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f008d96ef7e7319f7129146f5248a1dc376505e9d39ff950096d0322d5dbd08" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.307262 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-bbb3-account-create-2frqn" Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.309612 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"d0a05ea5748e5c0f888738c706e0ef9566711bd3c449ea3e49c606bc2599afb1"} Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.309674 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"b765875b0558b084ad69118fe44210f79623de6d959e05435a329cfeac851219"} Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.309688 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"36763225b7a46bafbf20f1b06f977e73b7a66eb2a50e5744facce41e0c20feb1"} Oct 04 11:06:57 crc kubenswrapper[4758]: I1004 11:06:57.309697 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"1da85af6ac2f7e956b4d723c710889da1c9cb0578fc66e469ed5aca8f74cf7fa"} Oct 04 11:06:57 crc kubenswrapper[4758]: E1004 11:06:57.398293 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab228585_a058_4d2a_aac4_6dd6bbf30852.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a3ed6de_8959_441c_bf24_025f5e7c185d.slice/crio-24551f8f4814d6826b0303eadf876af02f2ffe76a576dfc043c886127b46a8ce\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a3ed6de_8959_441c_bf24_025f5e7c185d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab228585_a058_4d2a_aac4_6dd6bbf30852.slice/crio-3377fb05556c966d97f90e5347b8f03c8eea9259bbe72a38f005cdcebacfbbb1\": RecentStats: unable to find data in memory cache]" Oct 04 11:07:01 crc kubenswrapper[4758]: I1004 11:07:01.250067 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:07:01 crc kubenswrapper[4758]: I1004 11:07:01.250560 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:07:01 crc kubenswrapper[4758]: I1004 11:07:01.348038 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"ee259ad5a389d1251c82dcafdfe37a7628c701836fefd58278d894c835248bbe"} Oct 04 11:07:01 crc kubenswrapper[4758]: I1004 11:07:01.348088 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"4347dbfe8447f357c238d93fa148cd177831a0299a2a873d4274f781781ae454"} Oct 04 11:07:01 crc kubenswrapper[4758]: I1004 11:07:01.348118 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"a09c1df8f724883389e8e3233e280faef25739b9cd9a1ec074c793d9cd409e82"} Oct 04 11:07:01 crc kubenswrapper[4758]: I1004 11:07:01.348132 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"6740b837d4042fd73a5d6190f6f0cf8ae515daf2cb6dfa1104b6b63ff3a90743"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375403 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"4395c6f004f66b24a0045cb182998e176207eba7f1018692996042f3ffa5a5c7"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375759 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"00d75dde3c3056b625883d7107303ae9417aba79535e73dcf7ab9a060064220a"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375769 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"1ae93d3ce9f993cbe716b0700873ca28d5cb27fa014331ecbc879598c0f771be"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375778 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"b57591994120b497f4c07b40d30680ba7956d6ee0cedba18ba5ae649afa9b3d6"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375787 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"512eacea9d05c98aae64a02cf33bacd8c4b7032e8fd5a1fbde3daf4a6ea89ffb"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"fac10c64511f2af4e9b55f1ee606c35cd1a78fb0da2e36cc2f30537fc8341e69"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.375804 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d3ed3476-2568-4c27-ac70-87d732389beb","Type":"ContainerStarted","Data":"b136b4ba95f91ab65657a034d054bce5547b953148bd5825fc9cfa4402dccd86"} Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.406682 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=51.792723325 podStartE2EDuration="59.406668177s" podCreationTimestamp="2025-10-04 11:06:04 +0000 UTC" firstStartedPulling="2025-10-04 11:06:54.594231541 +0000 UTC m=+991.886882430" lastFinishedPulling="2025-10-04 11:07:02.208176393 +0000 UTC m=+999.500827282" observedRunningTime="2025-10-04 11:07:03.403138631 +0000 UTC m=+1000.695789520" watchObservedRunningTime="2025-10-04 11:07:03.406668177 +0000 UTC m=+1000.699319066" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.657605 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dsjjc"] Oct 04 11:07:03 crc kubenswrapper[4758]: E1004 11:07:03.657949 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab228585-a058-4d2a-aac4-6dd6bbf30852" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.657961 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab228585-a058-4d2a-aac4-6dd6bbf30852" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: E1004 11:07:03.657984 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a3ed6de-8959-441c-bf24-025f5e7c185d" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.657990 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a3ed6de-8959-441c-bf24-025f5e7c185d" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: E1004 11:07:03.658000 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.658006 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.658170 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab228585-a058-4d2a-aac4-6dd6bbf30852" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.658197 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.658216 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a3ed6de-8959-441c-bf24-025f5e7c185d" containerName="mariadb-account-create" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.659000 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.660924 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.674631 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dsjjc"] Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.812733 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-svc\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.812841 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.812864 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8n8d\" (UniqueName: \"kubernetes.io/projected/2c897660-6520-4d3b-a424-51a12612f04c-kube-api-access-x8n8d\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.812895 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.812928 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-config\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.812954 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.915021 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-config\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.915121 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.915207 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-svc\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.915266 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.915291 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8n8d\" (UniqueName: \"kubernetes.io/projected/2c897660-6520-4d3b-a424-51a12612f04c-kube-api-access-x8n8d\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.915332 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.916382 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.916799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.917201 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.917274 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-config\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.919781 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-svc\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.934467 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8n8d\" (UniqueName: \"kubernetes.io/projected/2c897660-6520-4d3b-a424-51a12612f04c-kube-api-access-x8n8d\") pod \"dnsmasq-dns-764c5664d7-dsjjc\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:03 crc kubenswrapper[4758]: I1004 11:07:03.975011 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:04 crc kubenswrapper[4758]: I1004 11:07:04.389750 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dsjjc"] Oct 04 11:07:04 crc kubenswrapper[4758]: W1004 11:07:04.398176 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c897660_6520_4d3b_a424_51a12612f04c.slice/crio-74989e9dfb3e24c49c104f49cddc18f78743fe5028fdf60b858efce939a5ff07 WatchSource:0}: Error finding container 74989e9dfb3e24c49c104f49cddc18f78743fe5028fdf60b858efce939a5ff07: Status 404 returned error can't find the container with id 74989e9dfb3e24c49c104f49cddc18f78743fe5028fdf60b858efce939a5ff07 Oct 04 11:07:05 crc kubenswrapper[4758]: I1004 11:07:05.398741 4758 generic.go:334] "Generic (PLEG): container finished" podID="2c897660-6520-4d3b-a424-51a12612f04c" containerID="f945989308e4ae0cf187a2b8076d15b95ad9bcd934efb31de9b09a6568170b00" exitCode=0 Oct 04 11:07:05 crc kubenswrapper[4758]: I1004 11:07:05.398795 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" event={"ID":"2c897660-6520-4d3b-a424-51a12612f04c","Type":"ContainerDied","Data":"f945989308e4ae0cf187a2b8076d15b95ad9bcd934efb31de9b09a6568170b00"} Oct 04 11:07:05 crc kubenswrapper[4758]: I1004 11:07:05.398857 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" event={"ID":"2c897660-6520-4d3b-a424-51a12612f04c","Type":"ContainerStarted","Data":"74989e9dfb3e24c49c104f49cddc18f78743fe5028fdf60b858efce939a5ff07"} Oct 04 11:07:06 crc kubenswrapper[4758]: I1004 11:07:06.411870 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" event={"ID":"2c897660-6520-4d3b-a424-51a12612f04c","Type":"ContainerStarted","Data":"f3097b4bf3bd0cb71427d691c5a156d73277186bd12dbde749047c2c376dfa92"} Oct 04 11:07:06 crc kubenswrapper[4758]: I1004 11:07:06.412447 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:06 crc kubenswrapper[4758]: I1004 11:07:06.437814 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" podStartSLOduration=3.437788355 podStartE2EDuration="3.437788355s" podCreationTimestamp="2025-10-04 11:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:06.433116128 +0000 UTC m=+1003.725767017" watchObservedRunningTime="2025-10-04 11:07:06.437788355 +0000 UTC m=+1003.730439284" Oct 04 11:07:07 crc kubenswrapper[4758]: I1004 11:07:07.420707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-trp6p" event={"ID":"bfe10db6-4338-48ab-b744-18213468c240","Type":"ContainerStarted","Data":"733a2dbe8d4aa5ac353094a971ab1255ccba823e4bff633aaaaaea5e0500922b"} Oct 04 11:07:07 crc kubenswrapper[4758]: I1004 11:07:07.422552 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-crlgt" event={"ID":"157f102f-c0ec-4e03-97cb-9b45948c82ea","Type":"ContainerStarted","Data":"880004199ec2201e3d172dc100ed19d53d3e62e0e2b829dd48097b64ba29c9b8"} Oct 04 11:07:07 crc kubenswrapper[4758]: I1004 11:07:07.438214 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-trp6p" podStartSLOduration=1.681881856 podStartE2EDuration="33.438197545s" podCreationTimestamp="2025-10-04 11:06:34 +0000 UTC" firstStartedPulling="2025-10-04 11:06:35.02699749 +0000 UTC m=+972.319648379" lastFinishedPulling="2025-10-04 11:07:06.783313139 +0000 UTC m=+1004.075964068" observedRunningTime="2025-10-04 11:07:07.435263275 +0000 UTC m=+1004.727914204" watchObservedRunningTime="2025-10-04 11:07:07.438197545 +0000 UTC m=+1004.730848434" Oct 04 11:07:07 crc kubenswrapper[4758]: I1004 11:07:07.456029 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-crlgt" podStartSLOduration=1.986180084 podStartE2EDuration="34.456013958s" podCreationTimestamp="2025-10-04 11:06:33 +0000 UTC" firstStartedPulling="2025-10-04 11:06:34.427873988 +0000 UTC m=+971.720524877" lastFinishedPulling="2025-10-04 11:07:06.897707822 +0000 UTC m=+1004.190358751" observedRunningTime="2025-10-04 11:07:07.450273252 +0000 UTC m=+1004.742924141" watchObservedRunningTime="2025-10-04 11:07:07.456013958 +0000 UTC m=+1004.748664847" Oct 04 11:07:11 crc kubenswrapper[4758]: I1004 11:07:11.469616 4758 generic.go:334] "Generic (PLEG): container finished" podID="157f102f-c0ec-4e03-97cb-9b45948c82ea" containerID="880004199ec2201e3d172dc100ed19d53d3e62e0e2b829dd48097b64ba29c9b8" exitCode=0 Oct 04 11:07:11 crc kubenswrapper[4758]: I1004 11:07:11.469673 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-crlgt" event={"ID":"157f102f-c0ec-4e03-97cb-9b45948c82ea","Type":"ContainerDied","Data":"880004199ec2201e3d172dc100ed19d53d3e62e0e2b829dd48097b64ba29c9b8"} Oct 04 11:07:12 crc kubenswrapper[4758]: I1004 11:07:12.800493 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-crlgt" Oct 04 11:07:12 crc kubenswrapper[4758]: I1004 11:07:12.938079 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-config-data\") pod \"157f102f-c0ec-4e03-97cb-9b45948c82ea\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " Oct 04 11:07:12 crc kubenswrapper[4758]: I1004 11:07:12.938140 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgctw\" (UniqueName: \"kubernetes.io/projected/157f102f-c0ec-4e03-97cb-9b45948c82ea-kube-api-access-qgctw\") pod \"157f102f-c0ec-4e03-97cb-9b45948c82ea\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " Oct 04 11:07:12 crc kubenswrapper[4758]: I1004 11:07:12.938178 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-combined-ca-bundle\") pod \"157f102f-c0ec-4e03-97cb-9b45948c82ea\" (UID: \"157f102f-c0ec-4e03-97cb-9b45948c82ea\") " Oct 04 11:07:12 crc kubenswrapper[4758]: I1004 11:07:12.957144 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/157f102f-c0ec-4e03-97cb-9b45948c82ea-kube-api-access-qgctw" (OuterVolumeSpecName: "kube-api-access-qgctw") pod "157f102f-c0ec-4e03-97cb-9b45948c82ea" (UID: "157f102f-c0ec-4e03-97cb-9b45948c82ea"). InnerVolumeSpecName "kube-api-access-qgctw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:12 crc kubenswrapper[4758]: I1004 11:07:12.983955 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "157f102f-c0ec-4e03-97cb-9b45948c82ea" (UID: "157f102f-c0ec-4e03-97cb-9b45948c82ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.001645 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-config-data" (OuterVolumeSpecName: "config-data") pod "157f102f-c0ec-4e03-97cb-9b45948c82ea" (UID: "157f102f-c0ec-4e03-97cb-9b45948c82ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.040048 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.040245 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgctw\" (UniqueName: \"kubernetes.io/projected/157f102f-c0ec-4e03-97cb-9b45948c82ea-kube-api-access-qgctw\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.040258 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/157f102f-c0ec-4e03-97cb-9b45948c82ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.491714 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-crlgt" event={"ID":"157f102f-c0ec-4e03-97cb-9b45948c82ea","Type":"ContainerDied","Data":"93944d491d101fc3a574254df8f614dff30703fdc4720813814bf693379bb577"} Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.492015 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93944d491d101fc3a574254df8f614dff30703fdc4720813814bf693379bb577" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.491780 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-crlgt" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.809688 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2tr6m"] Oct 04 11:07:13 crc kubenswrapper[4758]: E1004 11:07:13.809985 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="157f102f-c0ec-4e03-97cb-9b45948c82ea" containerName="keystone-db-sync" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.809996 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="157f102f-c0ec-4e03-97cb-9b45948c82ea" containerName="keystone-db-sync" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.810196 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="157f102f-c0ec-4e03-97cb-9b45948c82ea" containerName="keystone-db-sync" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.810693 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.822485 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.822517 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.828637 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dsjjc"] Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.828881 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="dnsmasq-dns" containerID="cri-o://f3097b4bf3bd0cb71427d691c5a156d73277186bd12dbde749047c2c376dfa92" gracePeriod=10 Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.829001 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4n6pt" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.829014 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.830828 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.852301 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9r6c\" (UniqueName: \"kubernetes.io/projected/467014f5-f3df-466a-9d33-6eb3c70c174d-kube-api-access-w9r6c\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.852383 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-combined-ca-bundle\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.852432 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-scripts\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.852462 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-credential-keys\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.852489 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-config-data\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.852505 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-fernet-keys\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.954187 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-combined-ca-bundle\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.954260 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-scripts\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.954294 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-credential-keys\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.954312 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-config-data\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.954331 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-fernet-keys\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.954364 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9r6c\" (UniqueName: \"kubernetes.io/projected/467014f5-f3df-466a-9d33-6eb3c70c174d-kube-api-access-w9r6c\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.964264 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2tr6m"] Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.973231 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-scripts\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.975731 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-config-data\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.981260 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Oct 04 11:07:13 crc kubenswrapper[4758]: I1004 11:07:13.982732 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-fernet-keys\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.008500 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-credential-keys\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.008983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-combined-ca-bundle\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.016750 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7fgp7"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.019556 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9r6c\" (UniqueName: \"kubernetes.io/projected/467014f5-f3df-466a-9d33-6eb3c70c174d-kube-api-access-w9r6c\") pod \"keystone-bootstrap-2tr6m\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.021982 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.055786 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.055832 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.055866 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-svc\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.055893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.055923 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-config\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.055950 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pfdh\" (UniqueName: \"kubernetes.io/projected/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-kube-api-access-2pfdh\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.057750 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7fgp7"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.130156 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.148590 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8d48dff5f-86w22"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.150250 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158175 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pfdh\" (UniqueName: \"kubernetes.io/projected/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-kube-api-access-2pfdh\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158279 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158310 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158336 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-svc\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158356 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158385 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-config\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.158526 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8d48dff5f-86w22"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.159376 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-config\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.163438 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.165352 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.165547 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.165744 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-j6tch" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.165843 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.183640 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-svc\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.185065 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.186302 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.198877 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pfdh\" (UniqueName: \"kubernetes.io/projected/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-kube-api-access-2pfdh\") pod \"dnsmasq-dns-5959f8865f-7fgp7\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.202681 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.259684 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29n6g\" (UniqueName: \"kubernetes.io/projected/a34c59cb-b757-4d83-aca8-a68da66095b0-kube-api-access-29n6g\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.259721 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-scripts\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.259744 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-config-data\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.259778 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a34c59cb-b757-4d83-aca8-a68da66095b0-logs\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.259798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a34c59cb-b757-4d83-aca8-a68da66095b0-horizon-secret-key\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.267031 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-h9jzh"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.273057 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.279466 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jdw8z" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.279659 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.323751 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h9jzh"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.360686 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29n6g\" (UniqueName: \"kubernetes.io/projected/a34c59cb-b757-4d83-aca8-a68da66095b0-kube-api-access-29n6g\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.360721 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-scripts\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.360739 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-config-data\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.360772 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a34c59cb-b757-4d83-aca8-a68da66095b0-logs\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.360792 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a34c59cb-b757-4d83-aca8-a68da66095b0-horizon-secret-key\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.362719 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a34c59cb-b757-4d83-aca8-a68da66095b0-logs\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.363165 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-config-data\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.369153 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a34c59cb-b757-4d83-aca8-a68da66095b0-horizon-secret-key\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.375324 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-w9lbx"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.389049 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.404124 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.404284 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-np5w8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.404451 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.412692 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-w9lbx"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.413217 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29n6g\" (UniqueName: \"kubernetes.io/projected/a34c59cb-b757-4d83-aca8-a68da66095b0-kube-api-access-29n6g\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.422153 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qqgl8"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.423211 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.429240 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-scripts\") pod \"horizon-8d48dff5f-86w22\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.431052 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qqgl8"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.431571 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.431836 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pq4l8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.432039 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.487221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-config\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.487365 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plz7n\" (UniqueName: \"kubernetes.io/projected/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-kube-api-access-plz7n\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.487408 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-db-sync-config-data\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.487431 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-combined-ca-bundle\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.487498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-combined-ca-bundle\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.487560 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bwr2\" (UniqueName: \"kubernetes.io/projected/707c36c5-c69b-4d1d-8a0a-43891279b487-kube-api-access-2bwr2\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.517234 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7fgp7"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.532150 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-cd7cfc79c-4sgdt"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.533486 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.543289 4758 generic.go:334] "Generic (PLEG): container finished" podID="2c897660-6520-4d3b-a424-51a12612f04c" containerID="f3097b4bf3bd0cb71427d691c5a156d73277186bd12dbde749047c2c376dfa92" exitCode=0 Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.543411 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" event={"ID":"2c897660-6520-4d3b-a424-51a12612f04c","Type":"ContainerDied","Data":"f3097b4bf3bd0cb71427d691c5a156d73277186bd12dbde749047c2c376dfa92"} Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.558143 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.567349 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-mgdlb"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.568488 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.580894 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gp87p" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.580953 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.584775 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.588510 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b49gt\" (UniqueName: \"kubernetes.io/projected/d5a14d80-dbc2-4845-86c3-30c11ef4d085-kube-api-access-b49gt\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.588548 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-config-data\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.588571 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-scripts\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.588589 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r8j7\" (UniqueName: \"kubernetes.io/projected/41d2a43a-542a-40c5-884f-e6540082151e-kube-api-access-7r8j7\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.588614 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d2a43a-542a-40c5-884f-e6540082151e-etc-machine-id\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598181 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plz7n\" (UniqueName: \"kubernetes.io/projected/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-kube-api-access-plz7n\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598276 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-db-sync-config-data\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598295 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-config-data\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598342 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-combined-ca-bundle\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598390 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d5a14d80-dbc2-4845-86c3-30c11ef4d085-horizon-secret-key\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598409 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-logs\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598437 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxll9\" (UniqueName: \"kubernetes.io/projected/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-kube-api-access-jxll9\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598568 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-combined-ca-bundle\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598629 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-combined-ca-bundle\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598669 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-scripts\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598713 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-db-sync-config-data\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598729 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-combined-ca-bundle\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598754 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bwr2\" (UniqueName: \"kubernetes.io/projected/707c36c5-c69b-4d1d-8a0a-43891279b487-kube-api-access-2bwr2\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598788 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-config\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598809 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5a14d80-dbc2-4845-86c3-30c11ef4d085-logs\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598840 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-scripts\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.598860 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-config-data\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.604867 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-combined-ca-bundle\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.609388 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-db-sync-config-data\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.615190 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-combined-ca-bundle\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.616780 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-config\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.642079 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2pd6m"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.642656 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plz7n\" (UniqueName: \"kubernetes.io/projected/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-kube-api-access-plz7n\") pod \"barbican-db-sync-h9jzh\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.643226 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.643459 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.652311 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bwr2\" (UniqueName: \"kubernetes.io/projected/707c36c5-c69b-4d1d-8a0a-43891279b487-kube-api-access-2bwr2\") pod \"neutron-db-sync-qqgl8\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.663649 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.665629 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.667138 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.668064 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.688708 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mgdlb"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701019 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701050 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt49z\" (UniqueName: \"kubernetes.io/projected/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-kube-api-access-pt49z\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701073 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-combined-ca-bundle\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701115 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-scripts\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701138 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-db-sync-config-data\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701152 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-combined-ca-bundle\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701169 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-log-httpd\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701191 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-config-data\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701214 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5a14d80-dbc2-4845-86c3-30c11ef4d085-logs\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701244 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-scripts\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701261 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-config-data\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701282 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b49gt\" (UniqueName: \"kubernetes.io/projected/d5a14d80-dbc2-4845-86c3-30c11ef4d085-kube-api-access-b49gt\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701300 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-config-data\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701334 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-scripts\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701352 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r8j7\" (UniqueName: \"kubernetes.io/projected/41d2a43a-542a-40c5-884f-e6540082151e-kube-api-access-7r8j7\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701373 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.701394 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d2a43a-542a-40c5-884f-e6540082151e-etc-machine-id\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.704386 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5a14d80-dbc2-4845-86c3-30c11ef4d085-logs\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.703414 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708159 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-config-data\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708199 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708228 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d5a14d80-dbc2-4845-86c3-30c11ef4d085-horizon-secret-key\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708256 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-logs\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708270 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-scripts\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708321 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxll9\" (UniqueName: \"kubernetes.io/projected/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-kube-api-access-jxll9\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708342 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dn44\" (UniqueName: \"kubernetes.io/projected/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-kube-api-access-4dn44\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708387 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-config\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708411 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-run-httpd\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.708508 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d2a43a-542a-40c5-884f-e6540082151e-etc-machine-id\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.719080 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-logs\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.719495 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d5a14d80-dbc2-4845-86c3-30c11ef4d085-horizon-secret-key\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.720963 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-combined-ca-bundle\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.720974 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-scripts\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.730823 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-combined-ca-bundle\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.731833 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-config-data\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.734479 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-scripts\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.736331 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-db-sync-config-data\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.736807 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-scripts\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.741772 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-config-data\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.746143 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-config-data\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.750752 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cd7cfc79c-4sgdt"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.752970 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxll9\" (UniqueName: \"kubernetes.io/projected/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-kube-api-access-jxll9\") pod \"placement-db-sync-mgdlb\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.758973 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r8j7\" (UniqueName: \"kubernetes.io/projected/41d2a43a-542a-40c5-884f-e6540082151e-kube-api-access-7r8j7\") pod \"cinder-db-sync-w9lbx\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.763938 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.769024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b49gt\" (UniqueName: \"kubernetes.io/projected/d5a14d80-dbc2-4845-86c3-30c11ef4d085-kube-api-access-b49gt\") pod \"horizon-cd7cfc79c-4sgdt\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.774857 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2pd6m"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.790452 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810009 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810057 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810159 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810182 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-scripts\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810220 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dn44\" (UniqueName: \"kubernetes.io/projected/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-kube-api-access-4dn44\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810249 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-config\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810267 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-run-httpd\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810287 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810308 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt49z\" (UniqueName: \"kubernetes.io/projected/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-kube-api-access-pt49z\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810338 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-log-httpd\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.810359 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-config-data\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.814678 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-config-data\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.815811 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-config\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.816317 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-run-httpd\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.820531 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-log-httpd\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.821082 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.830418 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.832029 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-scripts\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.833117 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.834881 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.844732 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.844754 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.847947 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dn44\" (UniqueName: \"kubernetes.io/projected/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-kube-api-access-4dn44\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.859906 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt49z\" (UniqueName: \"kubernetes.io/projected/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-kube-api-access-pt49z\") pod \"ceilometer-0\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " pod="openstack/ceilometer-0" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.870610 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-2pd6m\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.925351 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.946462 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgdlb" Oct 04 11:07:14 crc kubenswrapper[4758]: I1004 11:07:14.956034 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:14.980117 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.005976 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.019522 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-config\") pod \"2c897660-6520-4d3b-a424-51a12612f04c\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.019558 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-svc\") pod \"2c897660-6520-4d3b-a424-51a12612f04c\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.019626 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-swift-storage-0\") pod \"2c897660-6520-4d3b-a424-51a12612f04c\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.019661 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8n8d\" (UniqueName: \"kubernetes.io/projected/2c897660-6520-4d3b-a424-51a12612f04c-kube-api-access-x8n8d\") pod \"2c897660-6520-4d3b-a424-51a12612f04c\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.019710 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-sb\") pod \"2c897660-6520-4d3b-a424-51a12612f04c\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.019776 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-nb\") pod \"2c897660-6520-4d3b-a424-51a12612f04c\" (UID: \"2c897660-6520-4d3b-a424-51a12612f04c\") " Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.071253 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c897660-6520-4d3b-a424-51a12612f04c-kube-api-access-x8n8d" (OuterVolumeSpecName: "kube-api-access-x8n8d") pod "2c897660-6520-4d3b-a424-51a12612f04c" (UID: "2c897660-6520-4d3b-a424-51a12612f04c"). InnerVolumeSpecName "kube-api-access-x8n8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.104164 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-config" (OuterVolumeSpecName: "config") pod "2c897660-6520-4d3b-a424-51a12612f04c" (UID: "2c897660-6520-4d3b-a424-51a12612f04c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.123124 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7fgp7"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.124212 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8n8d\" (UniqueName: \"kubernetes.io/projected/2c897660-6520-4d3b-a424-51a12612f04c-kube-api-access-x8n8d\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.124238 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:15 crc kubenswrapper[4758]: W1004 11:07:15.203022 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a43b127_b0e8_4ec7_ac3c_61816f7eb7a8.slice/crio-1ed0a3046fff1da29311b59e836d03ebb886a830a2105e6788fa9a90181abbbe WatchSource:0}: Error finding container 1ed0a3046fff1da29311b59e836d03ebb886a830a2105e6788fa9a90181abbbe: Status 404 returned error can't find the container with id 1ed0a3046fff1da29311b59e836d03ebb886a830a2105e6788fa9a90181abbbe Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.221401 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2c897660-6520-4d3b-a424-51a12612f04c" (UID: "2c897660-6520-4d3b-a424-51a12612f04c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.231523 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.236176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2c897660-6520-4d3b-a424-51a12612f04c" (UID: "2c897660-6520-4d3b-a424-51a12612f04c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.255547 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2c897660-6520-4d3b-a424-51a12612f04c" (UID: "2c897660-6520-4d3b-a424-51a12612f04c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.276627 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2c897660-6520-4d3b-a424-51a12612f04c" (UID: "2c897660-6520-4d3b-a424-51a12612f04c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.346272 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.346590 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.346672 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c897660-6520-4d3b-a424-51a12612f04c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.386663 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2tr6m"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.493584 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8d48dff5f-86w22"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.558599 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2tr6m" event={"ID":"467014f5-f3df-466a-9d33-6eb3c70c174d","Type":"ContainerStarted","Data":"f9e282d6a147baab0f457c165b0985fe79f9f776231ec2b53c92b3ffde385dd3"} Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.579583 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" event={"ID":"2c897660-6520-4d3b-a424-51a12612f04c","Type":"ContainerDied","Data":"74989e9dfb3e24c49c104f49cddc18f78743fe5028fdf60b858efce939a5ff07"} Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.579631 4758 scope.go:117] "RemoveContainer" containerID="f3097b4bf3bd0cb71427d691c5a156d73277186bd12dbde749047c2c376dfa92" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.579905 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-dsjjc" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.587194 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" event={"ID":"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8","Type":"ContainerStarted","Data":"297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12"} Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.587298 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" event={"ID":"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8","Type":"ContainerStarted","Data":"1ed0a3046fff1da29311b59e836d03ebb886a830a2105e6788fa9a90181abbbe"} Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.587382 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" podUID="5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" containerName="init" containerID="cri-o://297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12" gracePeriod=10 Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.593015 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8d48dff5f-86w22" event={"ID":"a34c59cb-b757-4d83-aca8-a68da66095b0","Type":"ContainerStarted","Data":"6d7660414ee9b26a3a3e1638c3d4115ce8133469222aaec07d6b261fbde26f97"} Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.674081 4758 scope.go:117] "RemoveContainer" containerID="f945989308e4ae0cf187a2b8076d15b95ad9bcd934efb31de9b09a6568170b00" Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.691145 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dsjjc"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.709642 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-dsjjc"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.730940 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-w9lbx"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.746964 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h9jzh"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.916186 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qqgl8"] Oct 04 11:07:15 crc kubenswrapper[4758]: I1004 11:07:15.946730 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cd7cfc79c-4sgdt"] Oct 04 11:07:15 crc kubenswrapper[4758]: W1004 11:07:15.977161 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5a14d80_dbc2_4845_86c3_30c11ef4d085.slice/crio-2c17831d31f1b0145d38defa514c6fd29920782772cf30c61fd5e1f7113ded00 WatchSource:0}: Error finding container 2c17831d31f1b0145d38defa514c6fd29920782772cf30c61fd5e1f7113ded00: Status 404 returned error can't find the container with id 2c17831d31f1b0145d38defa514c6fd29920782772cf30c61fd5e1f7113ded00 Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.053480 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-mgdlb"] Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.171660 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2pd6m"] Oct 04 11:07:16 crc kubenswrapper[4758]: W1004 11:07:16.196733 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9f151b2_a26c_4e4f_a728_a1ca1b7f58af.slice/crio-9945cc02733cb5988f659a282fab5a3e025896084a5694235e498c785a93ae1c WatchSource:0}: Error finding container 9945cc02733cb5988f659a282fab5a3e025896084a5694235e498c785a93ae1c: Status 404 returned error can't find the container with id 9945cc02733cb5988f659a282fab5a3e025896084a5694235e498c785a93ae1c Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.197164 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.391286 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.584624 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-nb\") pod \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.585181 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-swift-storage-0\") pod \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.585339 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-sb\") pod \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.585409 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-svc\") pod \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.585462 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pfdh\" (UniqueName: \"kubernetes.io/projected/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-kube-api-access-2pfdh\") pod \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.585936 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-config\") pod \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\" (UID: \"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8\") " Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.605514 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-kube-api-access-2pfdh" (OuterVolumeSpecName: "kube-api-access-2pfdh") pod "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" (UID: "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8"). InnerVolumeSpecName "kube-api-access-2pfdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.613548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" (UID: "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.616791 4758 generic.go:334] "Generic (PLEG): container finished" podID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerID="5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7" exitCode=0 Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.616852 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" event={"ID":"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af","Type":"ContainerDied","Data":"5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.616878 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" event={"ID":"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af","Type":"ContainerStarted","Data":"9945cc02733cb5988f659a282fab5a3e025896084a5694235e498c785a93ae1c"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.617789 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9jzh" event={"ID":"8ddb1196-e3aa-4fe4-8f93-e29c25a37664","Type":"ContainerStarted","Data":"b82e52602958ed05140872faca7122bfcf050062ff168dde2f56b85be5dcb311"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.618650 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cd7cfc79c-4sgdt" event={"ID":"d5a14d80-dbc2-4845-86c3-30c11ef4d085","Type":"ContainerStarted","Data":"2c17831d31f1b0145d38defa514c6fd29920782772cf30c61fd5e1f7113ded00"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.631345 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" (UID: "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.639217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w9lbx" event={"ID":"41d2a43a-542a-40c5-884f-e6540082151e","Type":"ContainerStarted","Data":"5d92bdb70f436860726f1bae62926df758cfd39693368881eec18cf68d3ff6d4"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.650501 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2tr6m" event={"ID":"467014f5-f3df-466a-9d33-6eb3c70c174d","Type":"ContainerStarted","Data":"7a8a8092571e2e857f73ce667c8831efeffaa296cc5f54f0c38f852454d3790a"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.653517 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" (UID: "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.657618 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" (UID: "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.666039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqgl8" event={"ID":"707c36c5-c69b-4d1d-8a0a-43891279b487","Type":"ContainerStarted","Data":"646b378ed44240903bc73e92a9185804f43c84628a92aa4752dac7debdc99783"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.666084 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqgl8" event={"ID":"707c36c5-c69b-4d1d-8a0a-43891279b487","Type":"ContainerStarted","Data":"dd2677ac41ee9ce7effae1cc01a7be93aec7f141878426c6ceb233170bf12250"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.669812 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerStarted","Data":"39b0eecef4325d0dd233eabe5622d7d5f3f0d1ca4d5b1cb933f9a6bce8923d07"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.671138 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgdlb" event={"ID":"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a","Type":"ContainerStarted","Data":"0a515c68c4ec77e06729358abbf715317e7dfda1da00673141cabdea7f01045f"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.678366 4758 generic.go:334] "Generic (PLEG): container finished" podID="5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" containerID="297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12" exitCode=0 Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.678403 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" event={"ID":"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8","Type":"ContainerDied","Data":"297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.678424 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" event={"ID":"5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8","Type":"ContainerDied","Data":"1ed0a3046fff1da29311b59e836d03ebb886a830a2105e6788fa9a90181abbbe"} Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.678440 4758 scope.go:117] "RemoveContainer" containerID="297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.678563 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-7fgp7" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.685355 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2tr6m" podStartSLOduration=3.6853424219999997 podStartE2EDuration="3.685342422s" podCreationTimestamp="2025-10-04 11:07:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:16.679169785 +0000 UTC m=+1013.971820674" watchObservedRunningTime="2025-10-04 11:07:16.685342422 +0000 UTC m=+1013.977993311" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.687928 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.687964 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.687990 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.688001 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.688010 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pfdh\" (UniqueName: \"kubernetes.io/projected/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-kube-api-access-2pfdh\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.689359 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-config" (OuterVolumeSpecName: "config") pod "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" (UID: "5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.700354 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qqgl8" podStartSLOduration=2.7003328890000002 podStartE2EDuration="2.700332889s" podCreationTimestamp="2025-10-04 11:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:16.695558379 +0000 UTC m=+1013.988209268" watchObservedRunningTime="2025-10-04 11:07:16.700332889 +0000 UTC m=+1013.992983778" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.724011 4758 scope.go:117] "RemoveContainer" containerID="297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12" Oct 04 11:07:16 crc kubenswrapper[4758]: E1004 11:07:16.724900 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12\": container with ID starting with 297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12 not found: ID does not exist" containerID="297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.724935 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12"} err="failed to get container status \"297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12\": rpc error: code = NotFound desc = could not find container \"297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12\": container with ID starting with 297c23ee7425b5d4b7fe2ac51964a189631bffb9eac362d2618cf2240add1b12 not found: ID does not exist" Oct 04 11:07:16 crc kubenswrapper[4758]: I1004 11:07:16.791645 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:16 crc kubenswrapper[4758]: E1004 11:07:16.856772 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T110716Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=bc5e38685c69e438103256e4ad466e078cd18d7f0cc945e73c57200dbfa8c065®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576936~hmac=40a92c93ff1e4eff5bc3365081404ffeb2a81fe24b9c8e51744c045dae1b4406\": EOF" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 11:07:16 crc kubenswrapper[4758]: E1004 11:07:16.856967 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndh576h56h659h559h5c6h5c7h654h5d6h58ch56fh5d4h69hd8h549h7fhcbhf4h87h67ch8bh677h64h5f7h696h57bh5ch5d9h5bbh664h7bh65bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b49gt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-cd7cfc79c-4sgdt_openstack(d5a14d80-dbc2-4845-86c3-30c11ef4d085): ErrImagePull: reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T110716Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=bc5e38685c69e438103256e4ad466e078cd18d7f0cc945e73c57200dbfa8c065®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576936~hmac=40a92c93ff1e4eff5bc3365081404ffeb2a81fe24b9c8e51744c045dae1b4406\": EOF" logger="UnhandledError" Oct 04 11:07:16 crc kubenswrapper[4758]: E1004 11:07:16.859532 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T110716Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=bc5e38685c69e438103256e4ad466e078cd18d7f0cc945e73c57200dbfa8c065®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576936~hmac=40a92c93ff1e4eff5bc3365081404ffeb2a81fe24b9c8e51744c045dae1b4406\\\": EOF\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-cd7cfc79c-4sgdt" podUID="d5a14d80-dbc2-4845-86c3-30c11ef4d085" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.116644 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7fgp7"] Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.120709 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-7fgp7"] Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.352516 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c897660-6520-4d3b-a424-51a12612f04c" path="/var/lib/kubelet/pods/2c897660-6520-4d3b-a424-51a12612f04c/volumes" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.353070 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" path="/var/lib/kubelet/pods/5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8/volumes" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.411969 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8d48dff5f-86w22"] Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.443795 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-655b4d654f-n47qb"] Oct 04 11:07:17 crc kubenswrapper[4758]: E1004 11:07:17.444496 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="dnsmasq-dns" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.444581 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="dnsmasq-dns" Oct 04 11:07:17 crc kubenswrapper[4758]: E1004 11:07:17.444648 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" containerName="init" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.444707 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" containerName="init" Oct 04 11:07:17 crc kubenswrapper[4758]: E1004 11:07:17.444770 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="init" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.444822 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="init" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.445030 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c897660-6520-4d3b-a424-51a12612f04c" containerName="dnsmasq-dns" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.445112 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a43b127-b0e8-4ec7-ac3c-61816f7eb7a8" containerName="init" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.446017 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.465398 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655b4d654f-n47qb"] Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.516639 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.612451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nz5c\" (UniqueName: \"kubernetes.io/projected/c245bc46-d74e-4129-a551-9a131a3b77bd-kube-api-access-4nz5c\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.612530 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-scripts\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.612586 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c245bc46-d74e-4129-a551-9a131a3b77bd-logs\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.612633 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-config-data\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.612657 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c245bc46-d74e-4129-a551-9a131a3b77bd-horizon-secret-key\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.701388 4758 generic.go:334] "Generic (PLEG): container finished" podID="bfe10db6-4338-48ab-b744-18213468c240" containerID="733a2dbe8d4aa5ac353094a971ab1255ccba823e4bff633aaaaaea5e0500922b" exitCode=0 Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.701441 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-trp6p" event={"ID":"bfe10db6-4338-48ab-b744-18213468c240","Type":"ContainerDied","Data":"733a2dbe8d4aa5ac353094a971ab1255ccba823e4bff633aaaaaea5e0500922b"} Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.706630 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" event={"ID":"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af","Type":"ContainerStarted","Data":"ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb"} Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.706660 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.713755 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c245bc46-d74e-4129-a551-9a131a3b77bd-logs\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.713813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-config-data\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.713846 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c245bc46-d74e-4129-a551-9a131a3b77bd-horizon-secret-key\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.713867 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nz5c\" (UniqueName: \"kubernetes.io/projected/c245bc46-d74e-4129-a551-9a131a3b77bd-kube-api-access-4nz5c\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.713917 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-scripts\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.714501 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-scripts\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.714695 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c245bc46-d74e-4129-a551-9a131a3b77bd-logs\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.715490 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-config-data\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: E1004 11:07:17.721060 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-cd7cfc79c-4sgdt" podUID="d5a14d80-dbc2-4845-86c3-30c11ef4d085" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.725461 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c245bc46-d74e-4129-a551-9a131a3b77bd-horizon-secret-key\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.736638 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nz5c\" (UniqueName: \"kubernetes.io/projected/c245bc46-d74e-4129-a551-9a131a3b77bd-kube-api-access-4nz5c\") pod \"horizon-655b4d654f-n47qb\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.783481 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" podStartSLOduration=3.7834644429999997 podStartE2EDuration="3.783464443s" podCreationTimestamp="2025-10-04 11:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:17.778908879 +0000 UTC m=+1015.071559758" watchObservedRunningTime="2025-10-04 11:07:17.783464443 +0000 UTC m=+1015.076115332" Oct 04 11:07:17 crc kubenswrapper[4758]: I1004 11:07:17.789708 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:18 crc kubenswrapper[4758]: I1004 11:07:18.463851 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655b4d654f-n47qb"] Oct 04 11:07:18 crc kubenswrapper[4758]: I1004 11:07:18.716550 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655b4d654f-n47qb" event={"ID":"c245bc46-d74e-4129-a551-9a131a3b77bd","Type":"ContainerStarted","Data":"7a3a4b82a84cf706c7999d9bbec3e5bfa993dbc55f7957d9ae0e9d3776b95bb5"} Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.161113 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-trp6p" Oct 04 11:07:19 crc kubenswrapper[4758]: E1004 11:07:19.265020 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T110718Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=e895d043e4ab0c64ec2179481817fd7c258320fc6f7c086b739118ee29413ec6®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576938~hmac=73410c2e1df98929369672ed0935b92916374a969515eb20a05fcc9ccc9cae00\": EOF" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 11:07:19 crc kubenswrapper[4758]: E1004 11:07:19.265211 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n688h5bch67bh58h5chbfh64ch55fh547h88h577h9fhf6h65fh6dh66dhbhd8hfhd9h646h8ch55dhfh6hfh55h57hb6h668h58h58fq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4nz5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-655b4d654f-n47qb_openstack(c245bc46-d74e-4129-a551-9a131a3b77bd): ErrImagePull: reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T110718Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=e895d043e4ab0c64ec2179481817fd7c258320fc6f7c086b739118ee29413ec6®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576938~hmac=73410c2e1df98929369672ed0935b92916374a969515eb20a05fcc9ccc9cae00\": EOF" logger="UnhandledError" Oct 04 11:07:19 crc kubenswrapper[4758]: E1004 11:07:19.267555 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T110718Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=e895d043e4ab0c64ec2179481817fd7c258320fc6f7c086b739118ee29413ec6®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576938~hmac=73410c2e1df98929369672ed0935b92916374a969515eb20a05fcc9ccc9cae00\\\": EOF\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-655b4d654f-n47qb" podUID="c245bc46-d74e-4129-a551-9a131a3b77bd" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.357518 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-combined-ca-bundle\") pod \"bfe10db6-4338-48ab-b744-18213468c240\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.357856 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82bdl\" (UniqueName: \"kubernetes.io/projected/bfe10db6-4338-48ab-b744-18213468c240-kube-api-access-82bdl\") pod \"bfe10db6-4338-48ab-b744-18213468c240\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.357902 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-db-sync-config-data\") pod \"bfe10db6-4338-48ab-b744-18213468c240\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.358260 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-config-data\") pod \"bfe10db6-4338-48ab-b744-18213468c240\" (UID: \"bfe10db6-4338-48ab-b744-18213468c240\") " Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.381428 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bfe10db6-4338-48ab-b744-18213468c240" (UID: "bfe10db6-4338-48ab-b744-18213468c240"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.387627 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe10db6-4338-48ab-b744-18213468c240-kube-api-access-82bdl" (OuterVolumeSpecName: "kube-api-access-82bdl") pod "bfe10db6-4338-48ab-b744-18213468c240" (UID: "bfe10db6-4338-48ab-b744-18213468c240"). InnerVolumeSpecName "kube-api-access-82bdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.400794 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe10db6-4338-48ab-b744-18213468c240" (UID: "bfe10db6-4338-48ab-b744-18213468c240"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.439406 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-config-data" (OuterVolumeSpecName: "config-data") pod "bfe10db6-4338-48ab-b744-18213468c240" (UID: "bfe10db6-4338-48ab-b744-18213468c240"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.465069 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.465130 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.465143 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82bdl\" (UniqueName: \"kubernetes.io/projected/bfe10db6-4338-48ab-b744-18213468c240-kube-api-access-82bdl\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.465152 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bfe10db6-4338-48ab-b744-18213468c240-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.728832 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-trp6p" Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.731141 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-trp6p" event={"ID":"bfe10db6-4338-48ab-b744-18213468c240","Type":"ContainerDied","Data":"028432c4df9435b2296627b31431737cd92de31914ae33e1620b16f5d821ea87"} Oct 04 11:07:19 crc kubenswrapper[4758]: I1004 11:07:19.731187 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="028432c4df9435b2296627b31431737cd92de31914ae33e1620b16f5d821ea87" Oct 04 11:07:19 crc kubenswrapper[4758]: E1004 11:07:19.733383 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-655b4d654f-n47qb" podUID="c245bc46-d74e-4129-a551-9a131a3b77bd" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.075261 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2pd6m"] Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.075477 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerName="dnsmasq-dns" containerID="cri-o://ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb" gracePeriod=10 Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.119784 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z8vnf"] Oct 04 11:07:20 crc kubenswrapper[4758]: E1004 11:07:20.122384 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe10db6-4338-48ab-b744-18213468c240" containerName="glance-db-sync" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.122406 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe10db6-4338-48ab-b744-18213468c240" containerName="glance-db-sync" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.122593 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe10db6-4338-48ab-b744-18213468c240" containerName="glance-db-sync" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.123397 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.151743 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z8vnf"] Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.205910 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-config\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.206283 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7bs2\" (UniqueName: \"kubernetes.io/projected/fedecb95-28eb-4fea-9b47-e01a16849726-kube-api-access-w7bs2\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.206333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.206361 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.206388 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.206422 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.308182 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.308317 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.308349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.308378 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.308406 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-config\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.308466 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7bs2\" (UniqueName: \"kubernetes.io/projected/fedecb95-28eb-4fea-9b47-e01a16849726-kube-api-access-w7bs2\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.309054 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.309511 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.309755 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-config\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.310183 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.310323 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.324814 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7bs2\" (UniqueName: \"kubernetes.io/projected/fedecb95-28eb-4fea-9b47-e01a16849726-kube-api-access-w7bs2\") pod \"dnsmasq-dns-785d8bcb8c-z8vnf\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.480245 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.685547 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.749920 4758 generic.go:334] "Generic (PLEG): container finished" podID="467014f5-f3df-466a-9d33-6eb3c70c174d" containerID="7a8a8092571e2e857f73ce667c8831efeffaa296cc5f54f0c38f852454d3790a" exitCode=0 Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.749981 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2tr6m" event={"ID":"467014f5-f3df-466a-9d33-6eb3c70c174d","Type":"ContainerDied","Data":"7a8a8092571e2e857f73ce667c8831efeffaa296cc5f54f0c38f852454d3790a"} Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.755720 4758 generic.go:334] "Generic (PLEG): container finished" podID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerID="ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb" exitCode=0 Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.755780 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" event={"ID":"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af","Type":"ContainerDied","Data":"ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb"} Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.755803 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" event={"ID":"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af","Type":"ContainerDied","Data":"9945cc02733cb5988f659a282fab5a3e025896084a5694235e498c785a93ae1c"} Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.755819 4758 scope.go:117] "RemoveContainer" containerID="ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.755874 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-2pd6m" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.819990 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dn44\" (UniqueName: \"kubernetes.io/projected/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-kube-api-access-4dn44\") pod \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.820115 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-sb\") pod \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.820201 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-swift-storage-0\") pod \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.820228 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-nb\") pod \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.820279 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-config\") pod \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.820321 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-svc\") pod \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\" (UID: \"e9f151b2-a26c-4e4f-a728-a1ca1b7f58af\") " Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.829164 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-kube-api-access-4dn44" (OuterVolumeSpecName: "kube-api-access-4dn44") pod "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" (UID: "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af"). InnerVolumeSpecName "kube-api-access-4dn44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.907238 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" (UID: "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.907426 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" (UID: "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.910280 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" (UID: "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.910550 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" (UID: "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.916956 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-config" (OuterVolumeSpecName: "config") pod "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" (UID: "e9f151b2-a26c-4e4f-a728-a1ca1b7f58af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.922649 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.922840 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.922850 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dn44\" (UniqueName: \"kubernetes.io/projected/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-kube-api-access-4dn44\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.922861 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.922928 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.922939 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.951539 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:20 crc kubenswrapper[4758]: E1004 11:07:20.951835 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerName="dnsmasq-dns" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.951850 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerName="dnsmasq-dns" Oct 04 11:07:20 crc kubenswrapper[4758]: E1004 11:07:20.951873 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerName="init" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.951880 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerName="init" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.952050 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" containerName="dnsmasq-dns" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.952962 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.956028 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8w5vq" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.959088 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.963853 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 11:07:20 crc kubenswrapper[4758]: I1004 11:07:20.971721 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.033267 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z8vnf"] Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.097289 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2pd6m"] Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.098916 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-2pd6m"] Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126213 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-config-data\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126315 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-scripts\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126386 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64q8b\" (UniqueName: \"kubernetes.io/projected/4c408503-6ea0-48fc-94ba-c419b0029434-kube-api-access-64q8b\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126418 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-logs\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126462 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.126520 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228193 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64q8b\" (UniqueName: \"kubernetes.io/projected/4c408503-6ea0-48fc-94ba-c419b0029434-kube-api-access-64q8b\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-logs\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228291 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228355 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-config-data\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228412 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-scripts\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.228973 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.229168 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.229282 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-logs\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.240271 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-scripts\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.241455 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.244832 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.246208 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.260255 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.260789 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-config-data\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.266746 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.293273 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64q8b\" (UniqueName: \"kubernetes.io/projected/4c408503-6ea0-48fc-94ba-c419b0029434-kube-api-access-64q8b\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.304508 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.374815 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9f151b2-a26c-4e4f-a728-a1ca1b7f58af" path="/var/lib/kubelet/pods/e9f151b2-a26c-4e4f-a728-a1ca1b7f58af/volumes" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432313 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432371 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432394 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-logs\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432418 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvtkv\" (UniqueName: \"kubernetes.io/projected/e1540d84-c216-4093-a344-5ac88db8c795-kube-api-access-bvtkv\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432514 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.432597 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534615 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534654 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-logs\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534679 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534702 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvtkv\" (UniqueName: \"kubernetes.io/projected/e1540d84-c216-4093-a344-5ac88db8c795-kube-api-access-bvtkv\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534752 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534811 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.534851 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.535526 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.535631 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.541066 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-logs\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.543389 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.543967 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.546737 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.559287 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvtkv\" (UniqueName: \"kubernetes.io/projected/e1540d84-c216-4093-a344-5ac88db8c795-kube-api-access-bvtkv\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.573411 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.574781 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:07:21 crc kubenswrapper[4758]: I1004 11:07:21.687848 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:24 crc kubenswrapper[4758]: W1004 11:07:24.708923 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfedecb95_28eb_4fea_9b47_e01a16849726.slice/crio-23134a3236d61ab6bdba6d02a6d0158f4c1f5c6c3ded800786dad96d7f40913e WatchSource:0}: Error finding container 23134a3236d61ab6bdba6d02a6d0158f4c1f5c6c3ded800786dad96d7f40913e: Status 404 returned error can't find the container with id 23134a3236d61ab6bdba6d02a6d0158f4c1f5c6c3ded800786dad96d7f40913e Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.720727 4758 scope.go:117] "RemoveContainer" containerID="5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.804987 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" event={"ID":"fedecb95-28eb-4fea-9b47-e01a16849726","Type":"ContainerStarted","Data":"23134a3236d61ab6bdba6d02a6d0158f4c1f5c6c3ded800786dad96d7f40913e"} Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.817549 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.821451 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2tr6m" event={"ID":"467014f5-f3df-466a-9d33-6eb3c70c174d","Type":"ContainerDied","Data":"f9e282d6a147baab0f457c165b0985fe79f9f776231ec2b53c92b3ffde385dd3"} Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.821491 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9e282d6a147baab0f457c165b0985fe79f9f776231ec2b53c92b3ffde385dd3" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.902768 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-credential-keys\") pod \"467014f5-f3df-466a-9d33-6eb3c70c174d\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.902853 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-scripts\") pod \"467014f5-f3df-466a-9d33-6eb3c70c174d\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.902877 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-combined-ca-bundle\") pod \"467014f5-f3df-466a-9d33-6eb3c70c174d\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.902931 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9r6c\" (UniqueName: \"kubernetes.io/projected/467014f5-f3df-466a-9d33-6eb3c70c174d-kube-api-access-w9r6c\") pod \"467014f5-f3df-466a-9d33-6eb3c70c174d\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.902990 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-fernet-keys\") pod \"467014f5-f3df-466a-9d33-6eb3c70c174d\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.903048 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-config-data\") pod \"467014f5-f3df-466a-9d33-6eb3c70c174d\" (UID: \"467014f5-f3df-466a-9d33-6eb3c70c174d\") " Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.911235 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/467014f5-f3df-466a-9d33-6eb3c70c174d-kube-api-access-w9r6c" (OuterVolumeSpecName: "kube-api-access-w9r6c") pod "467014f5-f3df-466a-9d33-6eb3c70c174d" (UID: "467014f5-f3df-466a-9d33-6eb3c70c174d"). InnerVolumeSpecName "kube-api-access-w9r6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.913662 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-scripts" (OuterVolumeSpecName: "scripts") pod "467014f5-f3df-466a-9d33-6eb3c70c174d" (UID: "467014f5-f3df-466a-9d33-6eb3c70c174d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.917675 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "467014f5-f3df-466a-9d33-6eb3c70c174d" (UID: "467014f5-f3df-466a-9d33-6eb3c70c174d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.920613 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "467014f5-f3df-466a-9d33-6eb3c70c174d" (UID: "467014f5-f3df-466a-9d33-6eb3c70c174d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.938259 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "467014f5-f3df-466a-9d33-6eb3c70c174d" (UID: "467014f5-f3df-466a-9d33-6eb3c70c174d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:24 crc kubenswrapper[4758]: I1004 11:07:24.941465 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-config-data" (OuterVolumeSpecName: "config-data") pod "467014f5-f3df-466a-9d33-6eb3c70c174d" (UID: "467014f5-f3df-466a-9d33-6eb3c70c174d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.005265 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.005303 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.005316 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9r6c\" (UniqueName: \"kubernetes.io/projected/467014f5-f3df-466a-9d33-6eb3c70c174d-kube-api-access-w9r6c\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.005327 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.005336 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.005343 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/467014f5-f3df-466a-9d33-6eb3c70c174d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.828810 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2tr6m" Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.984144 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2tr6m"] Oct 04 11:07:25 crc kubenswrapper[4758]: I1004 11:07:25.995971 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2tr6m"] Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.097660 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kpx5z"] Oct 04 11:07:26 crc kubenswrapper[4758]: E1004 11:07:26.098433 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="467014f5-f3df-466a-9d33-6eb3c70c174d" containerName="keystone-bootstrap" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.098520 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="467014f5-f3df-466a-9d33-6eb3c70c174d" containerName="keystone-bootstrap" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.098738 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="467014f5-f3df-466a-9d33-6eb3c70c174d" containerName="keystone-bootstrap" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.099259 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kpx5z"] Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.099392 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.128428 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.128624 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.128731 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4n6pt" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.128873 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.231135 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-fernet-keys\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.231222 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-credential-keys\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.231316 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-combined-ca-bundle\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.231342 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-scripts\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.231374 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-config-data\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.232470 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhcsd\" (UniqueName: \"kubernetes.io/projected/ca9af889-3123-43e5-b21b-c8cd9462563c-kube-api-access-qhcsd\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.333765 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-credential-keys\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.333882 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-combined-ca-bundle\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.333909 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-scripts\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.333945 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-config-data\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.333963 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhcsd\" (UniqueName: \"kubernetes.io/projected/ca9af889-3123-43e5-b21b-c8cd9462563c-kube-api-access-qhcsd\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.333984 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-fernet-keys\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.340056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-fernet-keys\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.341505 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-config-data\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.345169 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-combined-ca-bundle\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.357867 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-credential-keys\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.358666 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhcsd\" (UniqueName: \"kubernetes.io/projected/ca9af889-3123-43e5-b21b-c8cd9462563c-kube-api-access-qhcsd\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.361387 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-scripts\") pod \"keystone-bootstrap-kpx5z\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:26 crc kubenswrapper[4758]: I1004 11:07:26.457162 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:07:27 crc kubenswrapper[4758]: I1004 11:07:27.334817 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="467014f5-f3df-466a-9d33-6eb3c70c174d" path="/var/lib/kubelet/pods/467014f5-f3df-466a-9d33-6eb3c70c174d/volumes" Oct 04 11:07:27 crc kubenswrapper[4758]: I1004 11:07:27.357585 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:27 crc kubenswrapper[4758]: I1004 11:07:27.480202 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.194075 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cd7cfc79c-4sgdt"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.250520 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-549f9cdcb8-66q22"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.252467 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.255705 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.283018 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-549f9cdcb8-66q22"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.336561 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655b4d654f-n47qb"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370525 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-tls-certs\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-scripts\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370619 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-config-data\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370645 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phpw5\" (UniqueName: \"kubernetes.io/projected/55a162df-bf26-4c9f-9808-506db4c3bd01-kube-api-access-phpw5\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370702 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-combined-ca-bundle\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370731 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-secret-key\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.370761 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55a162df-bf26-4c9f-9808-506db4c3bd01-logs\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.383296 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c98cb656b-lczgq"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.384642 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.456603 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c98cb656b-lczgq"] Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474420 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-secret-key\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474488 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-horizon-secret-key\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474544 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55a162df-bf26-4c9f-9808-506db4c3bd01-logs\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474571 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-tls-certs\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474619 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-scripts\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474659 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61f721a8-00fb-446c-8e98-922827d2fc39-logs\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474690 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-config-data\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474736 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phpw5\" (UniqueName: \"kubernetes.io/projected/55a162df-bf26-4c9f-9808-506db4c3bd01-kube-api-access-phpw5\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474786 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x52ck\" (UniqueName: \"kubernetes.io/projected/61f721a8-00fb-446c-8e98-922827d2fc39-kube-api-access-x52ck\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474812 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-combined-ca-bundle\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474834 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61f721a8-00fb-446c-8e98-922827d2fc39-scripts\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-horizon-tls-certs\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474885 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61f721a8-00fb-446c-8e98-922827d2fc39-config-data\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.474953 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-combined-ca-bundle\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.477063 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55a162df-bf26-4c9f-9808-506db4c3bd01-logs\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.478606 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-scripts\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.485536 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-tls-certs\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.505340 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-config-data\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.521802 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-combined-ca-bundle\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.527734 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-secret-key\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.528456 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phpw5\" (UniqueName: \"kubernetes.io/projected/55a162df-bf26-4c9f-9808-506db4c3bd01-kube-api-access-phpw5\") pod \"horizon-549f9cdcb8-66q22\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.572374 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576337 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61f721a8-00fb-446c-8e98-922827d2fc39-logs\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576408 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x52ck\" (UniqueName: \"kubernetes.io/projected/61f721a8-00fb-446c-8e98-922827d2fc39-kube-api-access-x52ck\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576427 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-combined-ca-bundle\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576457 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61f721a8-00fb-446c-8e98-922827d2fc39-scripts\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576478 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-horizon-tls-certs\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576493 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61f721a8-00fb-446c-8e98-922827d2fc39-config-data\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.576562 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-horizon-secret-key\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.578276 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/61f721a8-00fb-446c-8e98-922827d2fc39-logs\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.579486 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61f721a8-00fb-446c-8e98-922827d2fc39-scripts\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.579624 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-horizon-secret-key\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.581123 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/61f721a8-00fb-446c-8e98-922827d2fc39-config-data\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.583362 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-combined-ca-bundle\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.585228 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/61f721a8-00fb-446c-8e98-922827d2fc39-horizon-tls-certs\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.608980 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x52ck\" (UniqueName: \"kubernetes.io/projected/61f721a8-00fb-446c-8e98-922827d2fc39-kube-api-access-x52ck\") pod \"horizon-5c98cb656b-lczgq\" (UID: \"61f721a8-00fb-446c-8e98-922827d2fc39\") " pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:28 crc kubenswrapper[4758]: I1004 11:07:28.754080 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.249508 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.249849 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.249894 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.250644 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b150cc50b08a3a5c33e13466357eafa5f32ab57cc3437792ce3dec5314cb49d"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.250687 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://2b150cc50b08a3a5c33e13466357eafa5f32ab57cc3437792ce3dec5314cb49d" gracePeriod=600 Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.926031 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="2b150cc50b08a3a5c33e13466357eafa5f32ab57cc3437792ce3dec5314cb49d" exitCode=0 Oct 04 11:07:31 crc kubenswrapper[4758]: I1004 11:07:31.926088 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"2b150cc50b08a3a5c33e13466357eafa5f32ab57cc3437792ce3dec5314cb49d"} Oct 04 11:07:34 crc kubenswrapper[4758]: E1004 11:07:34.665667 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 11:07:34 crc kubenswrapper[4758]: E1004 11:07:34.666356 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n675h97h684h566h56fh64bh5d4h5f9h7fhdh6fhdch5bbh674h658h57h59bh5f5h666hfch5f8h95h5ch59dh77h65bhbh659h548h6ch5d5hb5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-29n6g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-8d48dff5f-86w22_openstack(a34c59cb-b757-4d83-aca8-a68da66095b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:07:34 crc kubenswrapper[4758]: E1004 11:07:34.684364 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-8d48dff5f-86w22" podUID="a34c59cb-b757-4d83-aca8-a68da66095b0" Oct 04 11:07:36 crc kubenswrapper[4758]: E1004 11:07:36.469197 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 04 11:07:36 crc kubenswrapper[4758]: E1004 11:07:36.469898 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jxll9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-mgdlb_openstack(c471b6cd-fc95-4ce7-847f-3ea9cc76d74a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:07:36 crc kubenswrapper[4758]: E1004 11:07:36.471145 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-mgdlb" podUID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.540224 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.560477 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635159 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d5a14d80-dbc2-4845-86c3-30c11ef4d085-horizon-secret-key\") pod \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635199 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b49gt\" (UniqueName: \"kubernetes.io/projected/d5a14d80-dbc2-4845-86c3-30c11ef4d085-kube-api-access-b49gt\") pod \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635267 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-config-data\") pod \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635811 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-scripts" (OuterVolumeSpecName: "scripts") pod "d5a14d80-dbc2-4845-86c3-30c11ef4d085" (UID: "d5a14d80-dbc2-4845-86c3-30c11ef4d085"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635867 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-config-data" (OuterVolumeSpecName: "config-data") pod "d5a14d80-dbc2-4845-86c3-30c11ef4d085" (UID: "d5a14d80-dbc2-4845-86c3-30c11ef4d085"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635903 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-scripts\") pod \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.635934 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5a14d80-dbc2-4845-86c3-30c11ef4d085-logs\") pod \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\" (UID: \"d5a14d80-dbc2-4845-86c3-30c11ef4d085\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.636124 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5a14d80-dbc2-4845-86c3-30c11ef4d085-logs" (OuterVolumeSpecName: "logs") pod "d5a14d80-dbc2-4845-86c3-30c11ef4d085" (UID: "d5a14d80-dbc2-4845-86c3-30c11ef4d085"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.636426 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.636442 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5a14d80-dbc2-4845-86c3-30c11ef4d085-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.636451 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5a14d80-dbc2-4845-86c3-30c11ef4d085-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.641360 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5a14d80-dbc2-4845-86c3-30c11ef4d085-kube-api-access-b49gt" (OuterVolumeSpecName: "kube-api-access-b49gt") pod "d5a14d80-dbc2-4845-86c3-30c11ef4d085" (UID: "d5a14d80-dbc2-4845-86c3-30c11ef4d085"). InnerVolumeSpecName "kube-api-access-b49gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.649715 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5a14d80-dbc2-4845-86c3-30c11ef4d085-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d5a14d80-dbc2-4845-86c3-30c11ef4d085" (UID: "d5a14d80-dbc2-4845-86c3-30c11ef4d085"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.737680 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-scripts\") pod \"c245bc46-d74e-4129-a551-9a131a3b77bd\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.737790 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c245bc46-d74e-4129-a551-9a131a3b77bd-logs\") pod \"c245bc46-d74e-4129-a551-9a131a3b77bd\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.737823 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-config-data\") pod \"c245bc46-d74e-4129-a551-9a131a3b77bd\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.737908 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nz5c\" (UniqueName: \"kubernetes.io/projected/c245bc46-d74e-4129-a551-9a131a3b77bd-kube-api-access-4nz5c\") pod \"c245bc46-d74e-4129-a551-9a131a3b77bd\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.737982 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c245bc46-d74e-4129-a551-9a131a3b77bd-horizon-secret-key\") pod \"c245bc46-d74e-4129-a551-9a131a3b77bd\" (UID: \"c245bc46-d74e-4129-a551-9a131a3b77bd\") " Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.738209 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-scripts" (OuterVolumeSpecName: "scripts") pod "c245bc46-d74e-4129-a551-9a131a3b77bd" (UID: "c245bc46-d74e-4129-a551-9a131a3b77bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.738325 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d5a14d80-dbc2-4845-86c3-30c11ef4d085-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.738337 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b49gt\" (UniqueName: \"kubernetes.io/projected/d5a14d80-dbc2-4845-86c3-30c11ef4d085-kube-api-access-b49gt\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.738346 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.738940 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c245bc46-d74e-4129-a551-9a131a3b77bd-logs" (OuterVolumeSpecName: "logs") pod "c245bc46-d74e-4129-a551-9a131a3b77bd" (UID: "c245bc46-d74e-4129-a551-9a131a3b77bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.739274 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-config-data" (OuterVolumeSpecName: "config-data") pod "c245bc46-d74e-4129-a551-9a131a3b77bd" (UID: "c245bc46-d74e-4129-a551-9a131a3b77bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.742616 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c245bc46-d74e-4129-a551-9a131a3b77bd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c245bc46-d74e-4129-a551-9a131a3b77bd" (UID: "c245bc46-d74e-4129-a551-9a131a3b77bd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.748440 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c245bc46-d74e-4129-a551-9a131a3b77bd-kube-api-access-4nz5c" (OuterVolumeSpecName: "kube-api-access-4nz5c") pod "c245bc46-d74e-4129-a551-9a131a3b77bd" (UID: "c245bc46-d74e-4129-a551-9a131a3b77bd"). InnerVolumeSpecName "kube-api-access-4nz5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.840330 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c245bc46-d74e-4129-a551-9a131a3b77bd-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.840359 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c245bc46-d74e-4129-a551-9a131a3b77bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.840370 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nz5c\" (UniqueName: \"kubernetes.io/projected/c245bc46-d74e-4129-a551-9a131a3b77bd-kube-api-access-4nz5c\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.840380 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c245bc46-d74e-4129-a551-9a131a3b77bd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.972832 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655b4d654f-n47qb" event={"ID":"c245bc46-d74e-4129-a551-9a131a3b77bd","Type":"ContainerDied","Data":"7a3a4b82a84cf706c7999d9bbec3e5bfa993dbc55f7957d9ae0e9d3776b95bb5"} Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.972887 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655b4d654f-n47qb" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.980536 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cd7cfc79c-4sgdt" Oct 04 11:07:36 crc kubenswrapper[4758]: I1004 11:07:36.983357 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cd7cfc79c-4sgdt" event={"ID":"d5a14d80-dbc2-4845-86c3-30c11ef4d085","Type":"ContainerDied","Data":"2c17831d31f1b0145d38defa514c6fd29920782772cf30c61fd5e1f7113ded00"} Oct 04 11:07:36 crc kubenswrapper[4758]: E1004 11:07:36.983995 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-mgdlb" podUID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" Oct 04 11:07:37 crc kubenswrapper[4758]: I1004 11:07:37.046295 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cd7cfc79c-4sgdt"] Oct 04 11:07:37 crc kubenswrapper[4758]: I1004 11:07:37.060076 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-cd7cfc79c-4sgdt"] Oct 04 11:07:37 crc kubenswrapper[4758]: I1004 11:07:37.094029 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655b4d654f-n47qb"] Oct 04 11:07:37 crc kubenswrapper[4758]: I1004 11:07:37.099782 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-655b4d654f-n47qb"] Oct 04 11:07:37 crc kubenswrapper[4758]: I1004 11:07:37.336835 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c245bc46-d74e-4129-a551-9a131a3b77bd" path="/var/lib/kubelet/pods/c245bc46-d74e-4129-a551-9a131a3b77bd/volumes" Oct 04 11:07:37 crc kubenswrapper[4758]: I1004 11:07:37.337601 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5a14d80-dbc2-4845-86c3-30c11ef4d085" path="/var/lib/kubelet/pods/d5a14d80-dbc2-4845-86c3-30c11ef4d085/volumes" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.685409 4758 scope.go:117] "RemoveContainer" containerID="ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb" Oct 04 11:07:45 crc kubenswrapper[4758]: E1004 11:07:45.687584 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb\": container with ID starting with ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb not found: ID does not exist" containerID="ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.687642 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb"} err="failed to get container status \"ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb\": rpc error: code = NotFound desc = could not find container \"ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb\": container with ID starting with ba508b503b7ef670f7a2d34a7c74df132d2be4793938845edd85f96bd6aa95cb not found: ID does not exist" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.687673 4758 scope.go:117] "RemoveContainer" containerID="5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7" Oct 04 11:07:45 crc kubenswrapper[4758]: E1004 11:07:45.687937 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7\": container with ID starting with 5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7 not found: ID does not exist" containerID="5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.687971 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7"} err="failed to get container status \"5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7\": rpc error: code = NotFound desc = could not find container \"5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7\": container with ID starting with 5146a3fcc94097d277dbc9d37327fbc13529380d3c863fb83400578b0fd092c7 not found: ID does not exist" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.687990 4758 scope.go:117] "RemoveContainer" containerID="5c234c874dc874e8dd12e5faf6f9512499bc90758f95287be1214207762ac29f" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.789282 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.838359 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a34c59cb-b757-4d83-aca8-a68da66095b0-horizon-secret-key\") pod \"a34c59cb-b757-4d83-aca8-a68da66095b0\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.838425 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a34c59cb-b757-4d83-aca8-a68da66095b0-logs\") pod \"a34c59cb-b757-4d83-aca8-a68da66095b0\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.838469 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29n6g\" (UniqueName: \"kubernetes.io/projected/a34c59cb-b757-4d83-aca8-a68da66095b0-kube-api-access-29n6g\") pod \"a34c59cb-b757-4d83-aca8-a68da66095b0\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.838553 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-scripts\") pod \"a34c59cb-b757-4d83-aca8-a68da66095b0\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.838604 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-config-data\") pod \"a34c59cb-b757-4d83-aca8-a68da66095b0\" (UID: \"a34c59cb-b757-4d83-aca8-a68da66095b0\") " Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.839351 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a34c59cb-b757-4d83-aca8-a68da66095b0-logs" (OuterVolumeSpecName: "logs") pod "a34c59cb-b757-4d83-aca8-a68da66095b0" (UID: "a34c59cb-b757-4d83-aca8-a68da66095b0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.839429 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-scripts" (OuterVolumeSpecName: "scripts") pod "a34c59cb-b757-4d83-aca8-a68da66095b0" (UID: "a34c59cb-b757-4d83-aca8-a68da66095b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.839634 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-config-data" (OuterVolumeSpecName: "config-data") pod "a34c59cb-b757-4d83-aca8-a68da66095b0" (UID: "a34c59cb-b757-4d83-aca8-a68da66095b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.842171 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34c59cb-b757-4d83-aca8-a68da66095b0-kube-api-access-29n6g" (OuterVolumeSpecName: "kube-api-access-29n6g") pod "a34c59cb-b757-4d83-aca8-a68da66095b0" (UID: "a34c59cb-b757-4d83-aca8-a68da66095b0"). InnerVolumeSpecName "kube-api-access-29n6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.842677 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34c59cb-b757-4d83-aca8-a68da66095b0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a34c59cb-b757-4d83-aca8-a68da66095b0" (UID: "a34c59cb-b757-4d83-aca8-a68da66095b0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.940563 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.940923 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a34c59cb-b757-4d83-aca8-a68da66095b0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.941033 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a34c59cb-b757-4d83-aca8-a68da66095b0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.941133 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a34c59cb-b757-4d83-aca8-a68da66095b0-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:45 crc kubenswrapper[4758]: I1004 11:07:45.941203 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29n6g\" (UniqueName: \"kubernetes.io/projected/a34c59cb-b757-4d83-aca8-a68da66095b0-kube-api-access-29n6g\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:46 crc kubenswrapper[4758]: I1004 11:07:46.052775 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8d48dff5f-86w22" event={"ID":"a34c59cb-b757-4d83-aca8-a68da66095b0","Type":"ContainerDied","Data":"6d7660414ee9b26a3a3e1638c3d4115ce8133469222aaec07d6b261fbde26f97"} Oct 04 11:07:46 crc kubenswrapper[4758]: I1004 11:07:46.052858 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8d48dff5f-86w22" Oct 04 11:07:46 crc kubenswrapper[4758]: I1004 11:07:46.107588 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8d48dff5f-86w22"] Oct 04 11:07:46 crc kubenswrapper[4758]: I1004 11:07:46.113090 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8d48dff5f-86w22"] Oct 04 11:07:46 crc kubenswrapper[4758]: E1004 11:07:46.949356 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 04 11:07:46 crc kubenswrapper[4758]: E1004 11:07:46.949698 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7r8j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-w9lbx_openstack(41d2a43a-542a-40c5-884f-e6540082151e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:07:46 crc kubenswrapper[4758]: E1004 11:07:46.950970 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-w9lbx" podUID="41d2a43a-542a-40c5-884f-e6540082151e" Oct 04 11:07:47 crc kubenswrapper[4758]: E1004 11:07:47.076966 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-w9lbx" podUID="41d2a43a-542a-40c5-884f-e6540082151e" Oct 04 11:07:47 crc kubenswrapper[4758]: I1004 11:07:47.356842 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a34c59cb-b757-4d83-aca8-a68da66095b0" path="/var/lib/kubelet/pods/a34c59cb-b757-4d83-aca8-a68da66095b0/volumes" Oct 04 11:07:47 crc kubenswrapper[4758]: I1004 11:07:47.560812 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kpx5z"] Oct 04 11:07:47 crc kubenswrapper[4758]: I1004 11:07:47.610197 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-549f9cdcb8-66q22"] Oct 04 11:07:47 crc kubenswrapper[4758]: I1004 11:07:47.677513 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c98cb656b-lczgq"] Oct 04 11:07:47 crc kubenswrapper[4758]: I1004 11:07:47.845759 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.090903 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549f9cdcb8-66q22" event={"ID":"55a162df-bf26-4c9f-9808-506db4c3bd01","Type":"ContainerStarted","Data":"48a04c4673475bdee1e0e3b82de7092eb94182c76e58803f5cfcb28ab1c9cac9"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.092248 4758 generic.go:334] "Generic (PLEG): container finished" podID="fedecb95-28eb-4fea-9b47-e01a16849726" containerID="dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7" exitCode=0 Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.092282 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" event={"ID":"fedecb95-28eb-4fea-9b47-e01a16849726","Type":"ContainerDied","Data":"dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.095089 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerStarted","Data":"d568eaf29b4ed2db926b7cb81837f209219d5272453d804aec6a62e9c7db4624"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.101536 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"56399f3880bb0482a5458063cd6de09cb352f18aa4d98a39db29016b0272c352"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.121615 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9jzh" event={"ID":"8ddb1196-e3aa-4fe4-8f93-e29c25a37664","Type":"ContainerStarted","Data":"d86b51fa03ce7d44dda6d844d4981bf2c956086380b7941878bfb73fe9654feb"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.146152 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpx5z" event={"ID":"ca9af889-3123-43e5-b21b-c8cd9462563c","Type":"ContainerStarted","Data":"583899d86731878285b20ae19e2cfa6e62a111a204b272d7cbfe8499fb447beb"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.146427 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpx5z" event={"ID":"ca9af889-3123-43e5-b21b-c8cd9462563c","Type":"ContainerStarted","Data":"3065df987d657b221e00f09c4504c47fa7ee3a05e14fa2cc5ba5fbb0a644cf95"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.158124 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c98cb656b-lczgq" event={"ID":"61f721a8-00fb-446c-8e98-922827d2fc39","Type":"ContainerStarted","Data":"a12348795ea8f6842865e73e2d8051738d1f0dfbf2642dec81bf99262020b93e"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.159551 4758 generic.go:334] "Generic (PLEG): container finished" podID="707c36c5-c69b-4d1d-8a0a-43891279b487" containerID="646b378ed44240903bc73e92a9185804f43c84628a92aa4752dac7debdc99783" exitCode=0 Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.160211 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqgl8" event={"ID":"707c36c5-c69b-4d1d-8a0a-43891279b487","Type":"ContainerDied","Data":"646b378ed44240903bc73e92a9185804f43c84628a92aa4752dac7debdc99783"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.164938 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4c408503-6ea0-48fc-94ba-c419b0029434","Type":"ContainerStarted","Data":"44f5a1b2bec94d18f4d7bc4dfe17d5e9e6dcffa97b1ff81f85bc71eb6891d3c7"} Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.169277 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-h9jzh" podStartSLOduration=4.256018362 podStartE2EDuration="34.169256854s" podCreationTimestamp="2025-10-04 11:07:14 +0000 UTC" firstStartedPulling="2025-10-04 11:07:15.786237511 +0000 UTC m=+1013.078888400" lastFinishedPulling="2025-10-04 11:07:45.699476003 +0000 UTC m=+1042.992126892" observedRunningTime="2025-10-04 11:07:48.143225317 +0000 UTC m=+1045.435876206" watchObservedRunningTime="2025-10-04 11:07:48.169256854 +0000 UTC m=+1045.461907733" Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.210278 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kpx5z" podStartSLOduration=22.210260196 podStartE2EDuration="22.210260196s" podCreationTimestamp="2025-10-04 11:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:48.197717196 +0000 UTC m=+1045.490368085" watchObservedRunningTime="2025-10-04 11:07:48.210260196 +0000 UTC m=+1045.502911085" Oct 04 11:07:48 crc kubenswrapper[4758]: I1004 11:07:48.657060 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.179217 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4c408503-6ea0-48fc-94ba-c419b0029434","Type":"ContainerStarted","Data":"5c3283adc240eeca18497fbf63bc82cdf199712496ac99f75f6b510732da7df8"} Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.184563 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549f9cdcb8-66q22" event={"ID":"55a162df-bf26-4c9f-9808-506db4c3bd01","Type":"ContainerStarted","Data":"e10f41f31bbb4f5fd1672270aaecda9dd129769e4c72c63717d0592b51d67711"} Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.196635 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" event={"ID":"fedecb95-28eb-4fea-9b47-e01a16849726","Type":"ContainerStarted","Data":"27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe"} Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.196707 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.201579 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c98cb656b-lczgq" event={"ID":"61f721a8-00fb-446c-8e98-922827d2fc39","Type":"ContainerStarted","Data":"1879f1801937d4659da7d57131cbee347b49df5d138e811f68d0a8385f43bfe5"} Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.201606 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c98cb656b-lczgq" event={"ID":"61f721a8-00fb-446c-8e98-922827d2fc39","Type":"ContainerStarted","Data":"ee87ae7de6405a97f81cd1d83d069b1c6f7491bf82b09e80b70dcb61cc33a4d5"} Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.205137 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1540d84-c216-4093-a344-5ac88db8c795","Type":"ContainerStarted","Data":"0e9e8dbec906b3d28001cdec17b4c404ffa154f35510b6652a5bd04525304347"} Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.220036 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" podStartSLOduration=29.220015899 podStartE2EDuration="29.220015899s" podCreationTimestamp="2025-10-04 11:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:49.216806492 +0000 UTC m=+1046.509457391" watchObservedRunningTime="2025-10-04 11:07:49.220015899 +0000 UTC m=+1046.512666788" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.608826 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.633391 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c98cb656b-lczgq" podStartSLOduration=21.113375146 podStartE2EDuration="21.633366393s" podCreationTimestamp="2025-10-04 11:07:28 +0000 UTC" firstStartedPulling="2025-10-04 11:07:47.717245661 +0000 UTC m=+1045.009896550" lastFinishedPulling="2025-10-04 11:07:48.237236908 +0000 UTC m=+1045.529887797" observedRunningTime="2025-10-04 11:07:49.243810905 +0000 UTC m=+1046.536461824" watchObservedRunningTime="2025-10-04 11:07:49.633366393 +0000 UTC m=+1046.926017282" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.644571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-config\") pod \"707c36c5-c69b-4d1d-8a0a-43891279b487\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.644753 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bwr2\" (UniqueName: \"kubernetes.io/projected/707c36c5-c69b-4d1d-8a0a-43891279b487-kube-api-access-2bwr2\") pod \"707c36c5-c69b-4d1d-8a0a-43891279b487\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.644777 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-combined-ca-bundle\") pod \"707c36c5-c69b-4d1d-8a0a-43891279b487\" (UID: \"707c36c5-c69b-4d1d-8a0a-43891279b487\") " Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.654741 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/707c36c5-c69b-4d1d-8a0a-43891279b487-kube-api-access-2bwr2" (OuterVolumeSpecName: "kube-api-access-2bwr2") pod "707c36c5-c69b-4d1d-8a0a-43891279b487" (UID: "707c36c5-c69b-4d1d-8a0a-43891279b487"). InnerVolumeSpecName "kube-api-access-2bwr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.678465 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-config" (OuterVolumeSpecName: "config") pod "707c36c5-c69b-4d1d-8a0a-43891279b487" (UID: "707c36c5-c69b-4d1d-8a0a-43891279b487"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.683611 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "707c36c5-c69b-4d1d-8a0a-43891279b487" (UID: "707c36c5-c69b-4d1d-8a0a-43891279b487"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.752819 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.752852 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bwr2\" (UniqueName: \"kubernetes.io/projected/707c36c5-c69b-4d1d-8a0a-43891279b487-kube-api-access-2bwr2\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:49 crc kubenswrapper[4758]: I1004 11:07:49.752864 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/707c36c5-c69b-4d1d-8a0a-43891279b487-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.267967 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549f9cdcb8-66q22" event={"ID":"55a162df-bf26-4c9f-9808-506db4c3bd01","Type":"ContainerStarted","Data":"b7c2b7c398b965f1af57612a5dd546394a6cac04c79b30a09d3ac182e4e98488"} Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.277796 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqgl8" event={"ID":"707c36c5-c69b-4d1d-8a0a-43891279b487","Type":"ContainerDied","Data":"dd2677ac41ee9ce7effae1cc01a7be93aec7f141878426c6ceb233170bf12250"} Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.277840 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd2677ac41ee9ce7effae1cc01a7be93aec7f141878426c6ceb233170bf12250" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.273053 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqgl8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.289081 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1540d84-c216-4093-a344-5ac88db8c795","Type":"ContainerStarted","Data":"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357"} Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.303831 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerStarted","Data":"d729e35291e40a7db33b5296b8d19ad1aa2eadda21d1a98483af40a691588f69"} Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.366242 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-549f9cdcb8-66q22" podStartSLOduration=21.634475663 podStartE2EDuration="22.366219224s" podCreationTimestamp="2025-10-04 11:07:28 +0000 UTC" firstStartedPulling="2025-10-04 11:07:47.604473492 +0000 UTC m=+1044.897124381" lastFinishedPulling="2025-10-04 11:07:48.336217053 +0000 UTC m=+1045.628867942" observedRunningTime="2025-10-04 11:07:50.333318411 +0000 UTC m=+1047.625969300" watchObservedRunningTime="2025-10-04 11:07:50.366219224 +0000 UTC m=+1047.658870113" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.407250 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4c408503-6ea0-48fc-94ba-c419b0029434","Type":"ContainerStarted","Data":"6b769cbd55465c30af76382bc58c5578f3bbb6fa1236bb67296d657d617a5676"} Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.407418 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-log" containerID="cri-o://5c3283adc240eeca18497fbf63bc82cdf199712496ac99f75f6b510732da7df8" gracePeriod=30 Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.407840 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-httpd" containerID="cri-o://6b769cbd55465c30af76382bc58c5578f3bbb6fa1236bb67296d657d617a5676" gracePeriod=30 Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.437656 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z8vnf"] Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.464538 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgdlb" event={"ID":"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a","Type":"ContainerStarted","Data":"e521a77eb4ba65d1251271bec3d9d0e60da5c75272ab719171bf4e32e97ec946"} Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.491563 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=31.491545624 podStartE2EDuration="31.491545624s" podCreationTimestamp="2025-10-04 11:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:50.460354468 +0000 UTC m=+1047.753005367" watchObservedRunningTime="2025-10-04 11:07:50.491545624 +0000 UTC m=+1047.784196513" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.496978 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-4kkk9"] Oct 04 11:07:50 crc kubenswrapper[4758]: E1004 11:07:50.497321 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707c36c5-c69b-4d1d-8a0a-43891279b487" containerName="neutron-db-sync" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.497332 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="707c36c5-c69b-4d1d-8a0a-43891279b487" containerName="neutron-db-sync" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.497497 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="707c36c5-c69b-4d1d-8a0a-43891279b487" containerName="neutron-db-sync" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.498356 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.556600 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-4kkk9"] Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.583629 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-mgdlb" podStartSLOduration=3.315096926 podStartE2EDuration="36.58361137s" podCreationTimestamp="2025-10-04 11:07:14 +0000 UTC" firstStartedPulling="2025-10-04 11:07:16.054833998 +0000 UTC m=+1013.347484887" lastFinishedPulling="2025-10-04 11:07:49.323348442 +0000 UTC m=+1046.615999331" observedRunningTime="2025-10-04 11:07:50.547722308 +0000 UTC m=+1047.840373197" watchObservedRunningTime="2025-10-04 11:07:50.58361137 +0000 UTC m=+1047.876262259" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.615886 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.615935 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.615999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.616068 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-config\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.616157 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8x77\" (UniqueName: \"kubernetes.io/projected/5b91f490-3d53-4563-99ee-2f58cbfba4f2-kube-api-access-s8x77\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.616231 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-svc\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.627331 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66cf6c6588-6rzc8"] Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.628762 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.639789 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.641456 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-pq4l8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.641619 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.641775 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.665059 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66cf6c6588-6rzc8"] Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.717187 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-svc\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.718628 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.718662 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.718717 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.718779 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-config\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.718815 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8x77\" (UniqueName: \"kubernetes.io/projected/5b91f490-3d53-4563-99ee-2f58cbfba4f2-kube-api-access-s8x77\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.721682 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-svc\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.723822 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.724081 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.724543 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-config\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.727602 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.772807 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8x77\" (UniqueName: \"kubernetes.io/projected/5b91f490-3d53-4563-99ee-2f58cbfba4f2-kube-api-access-s8x77\") pod \"dnsmasq-dns-55f844cf75-4kkk9\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.821064 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-ovndb-tls-certs\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.821143 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-config\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.821180 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-httpd-config\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.821280 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-combined-ca-bundle\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.821304 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsngq\" (UniqueName: \"kubernetes.io/projected/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-kube-api-access-tsngq\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.862993 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.929396 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-ovndb-tls-certs\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.929446 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-config\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.929484 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-httpd-config\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.929627 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-combined-ca-bundle\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.929651 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsngq\" (UniqueName: \"kubernetes.io/projected/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-kube-api-access-tsngq\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.937746 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-ovndb-tls-certs\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.942912 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-config\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.949907 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-combined-ca-bundle\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.950474 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsngq\" (UniqueName: \"kubernetes.io/projected/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-kube-api-access-tsngq\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:50 crc kubenswrapper[4758]: I1004 11:07:50.962480 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-httpd-config\") pod \"neutron-66cf6c6588-6rzc8\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:50.998540 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.520699 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-4kkk9"] Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.539707 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1540d84-c216-4093-a344-5ac88db8c795","Type":"ContainerStarted","Data":"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d"} Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.539853 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-log" containerID="cri-o://1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357" gracePeriod=30 Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.540318 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-httpd" containerID="cri-o://815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d" gracePeriod=30 Oct 04 11:07:51 crc kubenswrapper[4758]: W1004 11:07:51.548818 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b91f490_3d53_4563_99ee_2f58cbfba4f2.slice/crio-71bed0efca6593802fb361aeeb957b817ffa5cb3b5ea63758f959610c16b262e WatchSource:0}: Error finding container 71bed0efca6593802fb361aeeb957b817ffa5cb3b5ea63758f959610c16b262e: Status 404 returned error can't find the container with id 71bed0efca6593802fb361aeeb957b817ffa5cb3b5ea63758f959610c16b262e Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.561671 4758 generic.go:334] "Generic (PLEG): container finished" podID="4c408503-6ea0-48fc-94ba-c419b0029434" containerID="6b769cbd55465c30af76382bc58c5578f3bbb6fa1236bb67296d657d617a5676" exitCode=0 Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.561707 4758 generic.go:334] "Generic (PLEG): container finished" podID="4c408503-6ea0-48fc-94ba-c419b0029434" containerID="5c3283adc240eeca18497fbf63bc82cdf199712496ac99f75f6b510732da7df8" exitCode=143 Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.561884 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" containerName="dnsmasq-dns" containerID="cri-o://27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe" gracePeriod=10 Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.562077 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4c408503-6ea0-48fc-94ba-c419b0029434","Type":"ContainerDied","Data":"6b769cbd55465c30af76382bc58c5578f3bbb6fa1236bb67296d657d617a5676"} Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.562117 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4c408503-6ea0-48fc-94ba-c419b0029434","Type":"ContainerDied","Data":"5c3283adc240eeca18497fbf63bc82cdf199712496ac99f75f6b510732da7df8"} Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.576299 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.576351 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.609509 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=31.609487690999998 podStartE2EDuration="31.609487691s" podCreationTimestamp="2025-10-04 11:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:51.608904575 +0000 UTC m=+1048.901555464" watchObservedRunningTime="2025-10-04 11:07:51.609487691 +0000 UTC m=+1048.902138580" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.690594 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.690882 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:51 crc kubenswrapper[4758]: I1004 11:07:51.763651 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66cf6c6588-6rzc8"] Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.000969 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158527 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-config-data\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158593 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158633 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-logs\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158665 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64q8b\" (UniqueName: \"kubernetes.io/projected/4c408503-6ea0-48fc-94ba-c419b0029434-kube-api-access-64q8b\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158746 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-combined-ca-bundle\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158784 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-httpd-run\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.158827 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-scripts\") pod \"4c408503-6ea0-48fc-94ba-c419b0029434\" (UID: \"4c408503-6ea0-48fc-94ba-c419b0029434\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.160574 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-logs" (OuterVolumeSpecName: "logs") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.160731 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.171263 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c408503-6ea0-48fc-94ba-c419b0029434-kube-api-access-64q8b" (OuterVolumeSpecName: "kube-api-access-64q8b") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "kube-api-access-64q8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.173334 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-scripts" (OuterVolumeSpecName: "scripts") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.214614 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.262251 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.262273 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.262293 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.262302 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c408503-6ea0-48fc-94ba-c419b0029434-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.262311 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64q8b\" (UniqueName: \"kubernetes.io/projected/4c408503-6ea0-48fc-94ba-c419b0029434-kube-api-access-64q8b\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.344712 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.356682 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.363870 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.364010 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.409547 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.415522 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.467556 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-scripts\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.467814 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-svc\") pod \"fedecb95-28eb-4fea-9b47-e01a16849726\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.467942 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-config\") pod \"fedecb95-28eb-4fea-9b47-e01a16849726\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468031 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-swift-storage-0\") pod \"fedecb95-28eb-4fea-9b47-e01a16849726\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468139 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-combined-ca-bundle\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468240 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-config-data\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-sb\") pod \"fedecb95-28eb-4fea-9b47-e01a16849726\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468450 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-httpd-run\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468547 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7bs2\" (UniqueName: \"kubernetes.io/projected/fedecb95-28eb-4fea-9b47-e01a16849726-kube-api-access-w7bs2\") pod \"fedecb95-28eb-4fea-9b47-e01a16849726\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468838 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-logs\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.468940 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.469025 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvtkv\" (UniqueName: \"kubernetes.io/projected/e1540d84-c216-4093-a344-5ac88db8c795-kube-api-access-bvtkv\") pod \"e1540d84-c216-4093-a344-5ac88db8c795\" (UID: \"e1540d84-c216-4093-a344-5ac88db8c795\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.469150 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-nb\") pod \"fedecb95-28eb-4fea-9b47-e01a16849726\" (UID: \"fedecb95-28eb-4fea-9b47-e01a16849726\") " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.483331 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-logs" (OuterVolumeSpecName: "logs") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.483546 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.514675 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1540d84-c216-4093-a344-5ac88db8c795-kube-api-access-bvtkv" (OuterVolumeSpecName: "kube-api-access-bvtkv") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "kube-api-access-bvtkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.514793 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-scripts" (OuterVolumeSpecName: "scripts") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.514878 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.515573 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fedecb95-28eb-4fea-9b47-e01a16849726-kube-api-access-w7bs2" (OuterVolumeSpecName: "kube-api-access-w7bs2") pod "fedecb95-28eb-4fea-9b47-e01a16849726" (UID: "fedecb95-28eb-4fea-9b47-e01a16849726"). InnerVolumeSpecName "kube-api-access-w7bs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.572888 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.572928 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvtkv\" (UniqueName: \"kubernetes.io/projected/e1540d84-c216-4093-a344-5ac88db8c795-kube-api-access-bvtkv\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.572939 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.572951 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.572963 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7bs2\" (UniqueName: \"kubernetes.io/projected/fedecb95-28eb-4fea-9b47-e01a16849726-kube-api-access-w7bs2\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.572982 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1540d84-c216-4093-a344-5ac88db8c795-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.600546 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cf6c6588-6rzc8" event={"ID":"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb","Type":"ContainerStarted","Data":"f44166449473b09958e24799008ef9e8bde5ae7007bb57a303f04d64a7025c01"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.647117 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.653818 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fedecb95-28eb-4fea-9b47-e01a16849726" (UID: "fedecb95-28eb-4fea-9b47-e01a16849726"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.658632 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fedecb95-28eb-4fea-9b47-e01a16849726" (UID: "fedecb95-28eb-4fea-9b47-e01a16849726"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.658704 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-config-data" (OuterVolumeSpecName: "config-data") pod "4c408503-6ea0-48fc-94ba-c419b0029434" (UID: "4c408503-6ea0-48fc-94ba-c419b0029434"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.668217 4758 generic.go:334] "Generic (PLEG): container finished" podID="fedecb95-28eb-4fea-9b47-e01a16849726" containerID="27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe" exitCode=0 Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.668358 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" event={"ID":"fedecb95-28eb-4fea-9b47-e01a16849726","Type":"ContainerDied","Data":"27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.668392 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" event={"ID":"fedecb95-28eb-4fea-9b47-e01a16849726","Type":"ContainerDied","Data":"23134a3236d61ab6bdba6d02a6d0158f4c1f5c6c3ded800786dad96d7f40913e"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.668429 4758 scope.go:117] "RemoveContainer" containerID="27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.669476 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-z8vnf" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.680718 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.680747 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.680759 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.680769 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c408503-6ea0-48fc-94ba-c419b0029434-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.698808 4758 generic.go:334] "Generic (PLEG): container finished" podID="e1540d84-c216-4093-a344-5ac88db8c795" containerID="815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d" exitCode=143 Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.698840 4758 generic.go:334] "Generic (PLEG): container finished" podID="e1540d84-c216-4093-a344-5ac88db8c795" containerID="1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357" exitCode=143 Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.698918 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1540d84-c216-4093-a344-5ac88db8c795","Type":"ContainerDied","Data":"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.698946 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1540d84-c216-4093-a344-5ac88db8c795","Type":"ContainerDied","Data":"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.698957 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e1540d84-c216-4093-a344-5ac88db8c795","Type":"ContainerDied","Data":"0e9e8dbec906b3d28001cdec17b4c404ffa154f35510b6652a5bd04525304347"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.699062 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.714579 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.730439 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fedecb95-28eb-4fea-9b47-e01a16849726" (UID: "fedecb95-28eb-4fea-9b47-e01a16849726"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.730830 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-config-data" (OuterVolumeSpecName: "config-data") pod "e1540d84-c216-4093-a344-5ac88db8c795" (UID: "e1540d84-c216-4093-a344-5ac88db8c795"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.732282 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-config" (OuterVolumeSpecName: "config") pod "fedecb95-28eb-4fea-9b47-e01a16849726" (UID: "fedecb95-28eb-4fea-9b47-e01a16849726"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.734298 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4c408503-6ea0-48fc-94ba-c419b0029434","Type":"ContainerDied","Data":"44f5a1b2bec94d18f4d7bc4dfe17d5e9e6dcffa97b1ff81f85bc71eb6891d3c7"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.734380 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.746638 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" event={"ID":"5b91f490-3d53-4563-99ee-2f58cbfba4f2","Type":"ContainerStarted","Data":"71bed0efca6593802fb361aeeb957b817ffa5cb3b5ea63758f959610c16b262e"} Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.784076 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.784330 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.784397 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.784459 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1540d84-c216-4093-a344-5ac88db8c795-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.796503 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fedecb95-28eb-4fea-9b47-e01a16849726" (UID: "fedecb95-28eb-4fea-9b47-e01a16849726"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.861915 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.873210 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.885967 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fedecb95-28eb-4fea-9b47-e01a16849726-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.904990 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:52 crc kubenswrapper[4758]: E1004 11:07:52.905404 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-httpd" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905423 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-httpd" Oct 04 11:07:52 crc kubenswrapper[4758]: E1004 11:07:52.905435 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-log" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905442 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-log" Oct 04 11:07:52 crc kubenswrapper[4758]: E1004 11:07:52.905449 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-httpd" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905455 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-httpd" Oct 04 11:07:52 crc kubenswrapper[4758]: E1004 11:07:52.905475 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" containerName="init" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905481 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" containerName="init" Oct 04 11:07:52 crc kubenswrapper[4758]: E1004 11:07:52.905508 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" containerName="dnsmasq-dns" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905514 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" containerName="dnsmasq-dns" Oct 04 11:07:52 crc kubenswrapper[4758]: E1004 11:07:52.905525 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-log" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905531 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-log" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905684 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-httpd" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905694 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-httpd" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905703 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" containerName="glance-log" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905713 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1540d84-c216-4093-a344-5ac88db8c795" containerName="glance-log" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.905721 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" containerName="dnsmasq-dns" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.906565 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.910953 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.912281 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 11:07:52 crc kubenswrapper[4758]: I1004 11:07:52.912717 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.014657 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z8vnf"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.017688 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-z8vnf"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.069625 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.075285 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090282 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vr4m\" (UniqueName: \"kubernetes.io/projected/3d059c9e-fa03-46dd-b113-de161b74ecea-kube-api-access-9vr4m\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090406 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090428 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090486 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090504 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090536 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-config-data\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090786 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-scripts\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.090814 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-logs\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.103786 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.113086 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.119769 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.121902 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.132146 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192094 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vr4m\" (UniqueName: \"kubernetes.io/projected/3d059c9e-fa03-46dd-b113-de161b74ecea-kube-api-access-9vr4m\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192241 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192294 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192309 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192355 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-config-data\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192374 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-scripts\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192388 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-logs\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.192981 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-logs\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.193910 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.194149 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.208417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-config-data\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.217043 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.225871 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-scripts\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.226724 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.233778 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vr4m\" (UniqueName: \"kubernetes.io/projected/3d059c9e-fa03-46dd-b113-de161b74ecea-kube-api-access-9vr4m\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.235202 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.247828 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295574 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295631 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295656 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295690 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295724 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295808 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.295833 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmwhc\" (UniqueName: \"kubernetes.io/projected/e6ad8010-4207-4d04-9765-5b81facf9cad-kube-api-access-dmwhc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.337893 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c408503-6ea0-48fc-94ba-c419b0029434" path="/var/lib/kubelet/pods/4c408503-6ea0-48fc-94ba-c419b0029434/volumes" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.338714 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1540d84-c216-4093-a344-5ac88db8c795" path="/var/lib/kubelet/pods/e1540d84-c216-4093-a344-5ac88db8c795/volumes" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.339320 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fedecb95-28eb-4fea-9b47-e01a16849726" path="/var/lib/kubelet/pods/fedecb95-28eb-4fea-9b47-e01a16849726/volumes" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400410 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400460 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmwhc\" (UniqueName: \"kubernetes.io/projected/e6ad8010-4207-4d04-9765-5b81facf9cad-kube-api-access-dmwhc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400515 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400617 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400651 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400672 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400689 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.400739 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.401580 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.402058 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.403912 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.407453 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.407527 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.409273 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.411402 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.417434 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmwhc\" (UniqueName: \"kubernetes.io/projected/e6ad8010-4207-4d04-9765-5b81facf9cad-kube-api-access-dmwhc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.437172 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:07:53 crc kubenswrapper[4758]: I1004 11:07:53.732077 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.279348 4758 scope.go:117] "RemoveContainer" containerID="dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.430063 4758 scope.go:117] "RemoveContainer" containerID="27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe" Oct 04 11:07:54 crc kubenswrapper[4758]: E1004 11:07:54.430928 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe\": container with ID starting with 27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe not found: ID does not exist" containerID="27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.431032 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe"} err="failed to get container status \"27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe\": rpc error: code = NotFound desc = could not find container \"27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe\": container with ID starting with 27f9deea3b93c439efe1133db3490bb8c57cbadea8d76e21bdabbd17abadc1fe not found: ID does not exist" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.431157 4758 scope.go:117] "RemoveContainer" containerID="dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7" Oct 04 11:07:54 crc kubenswrapper[4758]: E1004 11:07:54.436265 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7\": container with ID starting with dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7 not found: ID does not exist" containerID="dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.436312 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7"} err="failed to get container status \"dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7\": rpc error: code = NotFound desc = could not find container \"dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7\": container with ID starting with dba39931fc3a744e7a788695bb76030314f208f84c3249b35d5fea27dd668ed7 not found: ID does not exist" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.436340 4758 scope.go:117] "RemoveContainer" containerID="815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.538913 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d66f67fd7-kd9f7"] Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.545268 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.548022 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.548147 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.552511 4758 scope.go:117] "RemoveContainer" containerID="1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.602265 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d66f67fd7-kd9f7"] Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726398 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-ovndb-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726447 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-combined-ca-bundle\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jncsn\" (UniqueName: \"kubernetes.io/projected/f3b292da-54e8-4646-b70a-b2618eb4af4d-kube-api-access-jncsn\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726493 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-public-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726534 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-config\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726561 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-httpd-config\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.726582 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-internal-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.769160 4758 scope.go:117] "RemoveContainer" containerID="815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d" Oct 04 11:07:54 crc kubenswrapper[4758]: E1004 11:07:54.774303 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d\": container with ID starting with 815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d not found: ID does not exist" containerID="815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.774332 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d"} err="failed to get container status \"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d\": rpc error: code = NotFound desc = could not find container \"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d\": container with ID starting with 815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d not found: ID does not exist" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.774351 4758 scope.go:117] "RemoveContainer" containerID="1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357" Oct 04 11:07:54 crc kubenswrapper[4758]: E1004 11:07:54.774720 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357\": container with ID starting with 1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357 not found: ID does not exist" containerID="1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.774746 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357"} err="failed to get container status \"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357\": rpc error: code = NotFound desc = could not find container \"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357\": container with ID starting with 1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357 not found: ID does not exist" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.774758 4758 scope.go:117] "RemoveContainer" containerID="815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.775412 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d"} err="failed to get container status \"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d\": rpc error: code = NotFound desc = could not find container \"815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d\": container with ID starting with 815582f59f57c2f3e3bc22c3f6469f5a794ecb3ccc19a148615eaf54f60a5e0d not found: ID does not exist" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.775433 4758 scope.go:117] "RemoveContainer" containerID="1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.775853 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357"} err="failed to get container status \"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357\": rpc error: code = NotFound desc = could not find container \"1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357\": container with ID starting with 1dd846c9de5cfe0a18fdd98a37ec3c337080ccf206d5145648e0f10ed99d4357 not found: ID does not exist" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.775871 4758 scope.go:117] "RemoveContainer" containerID="6b769cbd55465c30af76382bc58c5578f3bbb6fa1236bb67296d657d617a5676" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.820866 4758 scope.go:117] "RemoveContainer" containerID="5c3283adc240eeca18497fbf63bc82cdf199712496ac99f75f6b510732da7df8" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828209 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-ovndb-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828258 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-combined-ca-bundle\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828283 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jncsn\" (UniqueName: \"kubernetes.io/projected/f3b292da-54e8-4646-b70a-b2618eb4af4d-kube-api-access-jncsn\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828300 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-public-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828340 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-config\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-httpd-config\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.828389 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-internal-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.843371 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-public-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.843749 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-config\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.843880 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-internal-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.844966 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-combined-ca-bundle\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.852896 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jncsn\" (UniqueName: \"kubernetes.io/projected/f3b292da-54e8-4646-b70a-b2618eb4af4d-kube-api-access-jncsn\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.868799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-httpd-config\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:54 crc kubenswrapper[4758]: I1004 11:07:54.869620 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3b292da-54e8-4646-b70a-b2618eb4af4d-ovndb-tls-certs\") pod \"neutron-7d66f67fd7-kd9f7\" (UID: \"f3b292da-54e8-4646-b70a-b2618eb4af4d\") " pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.075127 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.372169 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.500496 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.729017 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d66f67fd7-kd9f7"] Oct 04 11:07:55 crc kubenswrapper[4758]: W1004 11:07:55.742304 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3b292da_54e8_4646_b70a_b2618eb4af4d.slice/crio-7bbd76014e1b127dc82750a489d76346cb0a3ba68a6c5fb1e3422281ac022a23 WatchSource:0}: Error finding container 7bbd76014e1b127dc82750a489d76346cb0a3ba68a6c5fb1e3422281ac022a23: Status 404 returned error can't find the container with id 7bbd76014e1b127dc82750a489d76346cb0a3ba68a6c5fb1e3422281ac022a23 Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.833595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d059c9e-fa03-46dd-b113-de161b74ecea","Type":"ContainerStarted","Data":"ba398e1bf14f52ec117196a7eb831cf40141bea2d9dddba248ca36eadf6ca4fc"} Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.841356 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerID="a1837ae4301df85bb22d793762f5489552eac3aed4743a567b0bb8dd78761a2c" exitCode=0 Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.841417 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" event={"ID":"5b91f490-3d53-4563-99ee-2f58cbfba4f2","Type":"ContainerDied","Data":"a1837ae4301df85bb22d793762f5489552eac3aed4743a567b0bb8dd78761a2c"} Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.851323 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cf6c6588-6rzc8" event={"ID":"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb","Type":"ContainerStarted","Data":"b9ce32a8fce8271bdc3fbe67f0eee4af8f3430da4115e44584099b07cd3170c3"} Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.851394 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cf6c6588-6rzc8" event={"ID":"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb","Type":"ContainerStarted","Data":"7a57b136c7ec5a8825880144eddb8123052bc39b0a8532fc339e04bf6f56bd96"} Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.851514 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.868343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d66f67fd7-kd9f7" event={"ID":"f3b292da-54e8-4646-b70a-b2618eb4af4d","Type":"ContainerStarted","Data":"7bbd76014e1b127dc82750a489d76346cb0a3ba68a6c5fb1e3422281ac022a23"} Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.877753 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6ad8010-4207-4d04-9765-5b81facf9cad","Type":"ContainerStarted","Data":"6f624be0f16857efaae6834175be97e7e417e994431be5bf0a463a54a2dfef5a"} Oct 04 11:07:55 crc kubenswrapper[4758]: I1004 11:07:55.903974 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66cf6c6588-6rzc8" podStartSLOduration=5.903953972 podStartE2EDuration="5.903953972s" podCreationTimestamp="2025-10-04 11:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:07:55.887399973 +0000 UTC m=+1053.180050862" watchObservedRunningTime="2025-10-04 11:07:55.903953972 +0000 UTC m=+1053.196604861" Oct 04 11:07:56 crc kubenswrapper[4758]: I1004 11:07:56.891219 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d059c9e-fa03-46dd-b113-de161b74ecea","Type":"ContainerStarted","Data":"9c1625a48956f88788991172d2be04f15d991d81c49ef25582359fdbff2fb32d"} Oct 04 11:07:56 crc kubenswrapper[4758]: I1004 11:07:56.893641 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6ad8010-4207-4d04-9765-5b81facf9cad","Type":"ContainerStarted","Data":"190b414813fbd25927e7ab3d30a73c8c1449cd4df4fb8e8945f3c6f5e3b9cbfb"} Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.573955 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.574304 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.755168 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.755362 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.765366 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c98cb656b-lczgq" podUID="61f721a8-00fb-446c-8e98-922827d2fc39" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.935676 4758 generic.go:334] "Generic (PLEG): container finished" podID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" containerID="e521a77eb4ba65d1251271bec3d9d0e60da5c75272ab719171bf4e32e97ec946" exitCode=0 Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.935730 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgdlb" event={"ID":"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a","Type":"ContainerDied","Data":"e521a77eb4ba65d1251271bec3d9d0e60da5c75272ab719171bf4e32e97ec946"} Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.937585 4758 generic.go:334] "Generic (PLEG): container finished" podID="ca9af889-3123-43e5-b21b-c8cd9462563c" containerID="583899d86731878285b20ae19e2cfa6e62a111a204b272d7cbfe8499fb447beb" exitCode=0 Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.937637 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpx5z" event={"ID":"ca9af889-3123-43e5-b21b-c8cd9462563c","Type":"ContainerDied","Data":"583899d86731878285b20ae19e2cfa6e62a111a204b272d7cbfe8499fb447beb"} Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.951282 4758 generic.go:334] "Generic (PLEG): container finished" podID="8ddb1196-e3aa-4fe4-8f93-e29c25a37664" containerID="d86b51fa03ce7d44dda6d844d4981bf2c956086380b7941878bfb73fe9654feb" exitCode=0 Oct 04 11:07:58 crc kubenswrapper[4758]: I1004 11:07:58.951817 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9jzh" event={"ID":"8ddb1196-e3aa-4fe4-8f93-e29c25a37664","Type":"ContainerDied","Data":"d86b51fa03ce7d44dda6d844d4981bf2c956086380b7941878bfb73fe9654feb"} Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.598840 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.603398 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgdlb" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676517 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-scripts\") pod \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676553 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-combined-ca-bundle\") pod \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676602 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-logs\") pod \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676661 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-config-data\") pod \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676680 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plz7n\" (UniqueName: \"kubernetes.io/projected/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-kube-api-access-plz7n\") pod \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676772 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-combined-ca-bundle\") pod \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676791 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxll9\" (UniqueName: \"kubernetes.io/projected/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-kube-api-access-jxll9\") pod \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\" (UID: \"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.676816 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-db-sync-config-data\") pod \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\" (UID: \"8ddb1196-e3aa-4fe4-8f93-e29c25a37664\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.699237 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8ddb1196-e3aa-4fe4-8f93-e29c25a37664" (UID: "8ddb1196-e3aa-4fe4-8f93-e29c25a37664"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.710294 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-kube-api-access-plz7n" (OuterVolumeSpecName: "kube-api-access-plz7n") pod "8ddb1196-e3aa-4fe4-8f93-e29c25a37664" (UID: "8ddb1196-e3aa-4fe4-8f93-e29c25a37664"). InnerVolumeSpecName "kube-api-access-plz7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.711348 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-logs" (OuterVolumeSpecName: "logs") pod "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" (UID: "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.732270 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-kube-api-access-jxll9" (OuterVolumeSpecName: "kube-api-access-jxll9") pod "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" (UID: "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a"). InnerVolumeSpecName "kube-api-access-jxll9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.733236 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-scripts" (OuterVolumeSpecName: "scripts") pod "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" (UID: "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.762272 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-config-data" (OuterVolumeSpecName: "config-data") pod "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" (UID: "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.778773 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.778807 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.778815 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.778824 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plz7n\" (UniqueName: \"kubernetes.io/projected/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-kube-api-access-plz7n\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.778835 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxll9\" (UniqueName: \"kubernetes.io/projected/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-kube-api-access-jxll9\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.778843 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.797466 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" (UID: "c471b6cd-fc95-4ce7-847f-3ea9cc76d74a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.802943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ddb1196-e3aa-4fe4-8f93-e29c25a37664" (UID: "8ddb1196-e3aa-4fe4-8f93-e29c25a37664"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.824952 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.880415 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-fernet-keys\") pod \"ca9af889-3123-43e5-b21b-c8cd9462563c\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.880477 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-credential-keys\") pod \"ca9af889-3123-43e5-b21b-c8cd9462563c\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.880536 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-combined-ca-bundle\") pod \"ca9af889-3123-43e5-b21b-c8cd9462563c\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.880592 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-config-data\") pod \"ca9af889-3123-43e5-b21b-c8cd9462563c\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.880630 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-scripts\") pod \"ca9af889-3123-43e5-b21b-c8cd9462563c\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.880794 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhcsd\" (UniqueName: \"kubernetes.io/projected/ca9af889-3123-43e5-b21b-c8cd9462563c-kube-api-access-qhcsd\") pod \"ca9af889-3123-43e5-b21b-c8cd9462563c\" (UID: \"ca9af889-3123-43e5-b21b-c8cd9462563c\") " Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.881190 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddb1196-e3aa-4fe4-8f93-e29c25a37664-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.881203 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.885550 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9af889-3123-43e5-b21b-c8cd9462563c-kube-api-access-qhcsd" (OuterVolumeSpecName: "kube-api-access-qhcsd") pod "ca9af889-3123-43e5-b21b-c8cd9462563c" (UID: "ca9af889-3123-43e5-b21b-c8cd9462563c"). InnerVolumeSpecName "kube-api-access-qhcsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.895177 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ca9af889-3123-43e5-b21b-c8cd9462563c" (UID: "ca9af889-3123-43e5-b21b-c8cd9462563c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.929283 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ca9af889-3123-43e5-b21b-c8cd9462563c" (UID: "ca9af889-3123-43e5-b21b-c8cd9462563c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.945821 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-scripts" (OuterVolumeSpecName: "scripts") pod "ca9af889-3123-43e5-b21b-c8cd9462563c" (UID: "ca9af889-3123-43e5-b21b-c8cd9462563c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.977886 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpx5z" event={"ID":"ca9af889-3123-43e5-b21b-c8cd9462563c","Type":"ContainerDied","Data":"3065df987d657b221e00f09c4504c47fa7ee3a05e14fa2cc5ba5fbb0a644cf95"} Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.978212 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3065df987d657b221e00f09c4504c47fa7ee3a05e14fa2cc5ba5fbb0a644cf95" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.978275 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpx5z" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.986642 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.986675 4758 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.986686 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.986694 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhcsd\" (UniqueName: \"kubernetes.io/projected/ca9af889-3123-43e5-b21b-c8cd9462563c-kube-api-access-qhcsd\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.988150 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9jzh" event={"ID":"8ddb1196-e3aa-4fe4-8f93-e29c25a37664","Type":"ContainerDied","Data":"b82e52602958ed05140872faca7122bfcf050062ff168dde2f56b85be5dcb311"} Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.988182 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b82e52602958ed05140872faca7122bfcf050062ff168dde2f56b85be5dcb311" Oct 04 11:08:00 crc kubenswrapper[4758]: I1004 11:08:00.988246 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9jzh" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.009904 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-mgdlb" event={"ID":"c471b6cd-fc95-4ce7-847f-3ea9cc76d74a","Type":"ContainerDied","Data":"0a515c68c4ec77e06729358abbf715317e7dfda1da00673141cabdea7f01045f"} Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.009938 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a515c68c4ec77e06729358abbf715317e7dfda1da00673141cabdea7f01045f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.010008 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-mgdlb" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.016661 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca9af889-3123-43e5-b21b-c8cd9462563c" (UID: "ca9af889-3123-43e5-b21b-c8cd9462563c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.056386 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-config-data" (OuterVolumeSpecName: "config-data") pod "ca9af889-3123-43e5-b21b-c8cd9462563c" (UID: "ca9af889-3123-43e5-b21b-c8cd9462563c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.117595 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.117803 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca9af889-3123-43e5-b21b-c8cd9462563c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.127660 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8488664d88-ghdnd"] Oct 04 11:08:01 crc kubenswrapper[4758]: E1004 11:08:01.148960 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddb1196-e3aa-4fe4-8f93-e29c25a37664" containerName="barbican-db-sync" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.156807 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddb1196-e3aa-4fe4-8f93-e29c25a37664" containerName="barbican-db-sync" Oct 04 11:08:01 crc kubenswrapper[4758]: E1004 11:08:01.157130 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9af889-3123-43e5-b21b-c8cd9462563c" containerName="keystone-bootstrap" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.160964 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9af889-3123-43e5-b21b-c8cd9462563c" containerName="keystone-bootstrap" Oct 04 11:08:01 crc kubenswrapper[4758]: E1004 11:08:01.161055 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" containerName="placement-db-sync" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.161066 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" containerName="placement-db-sync" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.187876 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" containerName="placement-db-sync" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.187909 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddb1196-e3aa-4fe4-8f93-e29c25a37664" containerName="barbican-db-sync" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.187942 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9af889-3123-43e5-b21b-c8cd9462563c" containerName="keystone-bootstrap" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.189317 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8488664d88-ghdnd"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.189418 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.198661 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.199524 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.199714 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gp87p" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.199750 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.211113 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.255900 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ccf54cd5-74s2f"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.259161 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.266666 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.266710 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.307204 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ccf54cd5-74s2f"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.354712 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-logs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.354941 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-credential-keys\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355017 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-scripts\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355094 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-config-data\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355198 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-internal-tls-certs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355282 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqmvm\" (UniqueName: \"kubernetes.io/projected/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-kube-api-access-jqmvm\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355368 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-fernet-keys\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355465 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-internal-tls-certs\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355572 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkj2r\" (UniqueName: \"kubernetes.io/projected/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-kube-api-access-bkj2r\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355645 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-public-tls-certs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355782 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-scripts\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355874 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-combined-ca-bundle\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.355965 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-config-data\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.356039 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-public-tls-certs\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.356157 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-combined-ca-bundle\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.437942 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5645f9f854-tfsnf"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.439478 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.451621 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.452255 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.452486 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jdw8z" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459472 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-combined-ca-bundle\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459522 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-logs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-credential-keys\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459561 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-scripts\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459595 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-config-data\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-internal-tls-certs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459654 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqmvm\" (UniqueName: \"kubernetes.io/projected/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-kube-api-access-jqmvm\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459689 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-fernet-keys\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459715 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-internal-tls-certs\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459738 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkj2r\" (UniqueName: \"kubernetes.io/projected/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-kube-api-access-bkj2r\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459761 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-public-tls-certs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-scripts\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459823 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-combined-ca-bundle\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459860 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-config-data\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.459879 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-public-tls-certs\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.461777 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-logs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.472014 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-combined-ca-bundle\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.484330 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-internal-tls-certs\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.491401 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-public-tls-certs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.502216 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-75574c7dff-p5nq5"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.503658 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.518088 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-combined-ca-bundle\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.518876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-scripts\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.547813 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-config-data\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.553748 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-fernet-keys\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.554131 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-credential-keys\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.554297 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.554409 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-scripts\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.554888 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-config-data\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.555294 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-internal-tls-certs\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.556148 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-public-tls-certs\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.556523 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkj2r\" (UniqueName: \"kubernetes.io/projected/6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce-kube-api-access-bkj2r\") pod \"keystone-ccf54cd5-74s2f\" (UID: \"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce\") " pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.560654 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqmvm\" (UniqueName: \"kubernetes.io/projected/96e4bc74-27b8-4c61-85c5-d56589b9e4bf-kube-api-access-jqmvm\") pod \"placement-8488664d88-ghdnd\" (UID: \"96e4bc74-27b8-4c61-85c5-d56589b9e4bf\") " pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.561575 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-config-data-custom\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.561643 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-combined-ca-bundle\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.561675 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-config-data\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.561708 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-logs\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.561729 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mx8f\" (UniqueName: \"kubernetes.io/projected/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-kube-api-access-8mx8f\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.576322 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5645f9f854-tfsnf"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.669544 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671385 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84de7aef-2467-44a1-ae92-c22d87035c0d-logs\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671458 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-combined-ca-bundle\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671493 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-combined-ca-bundle\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671515 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-config-data\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-config-data\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671584 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-logs\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671612 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mx8f\" (UniqueName: \"kubernetes.io/projected/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-kube-api-access-8mx8f\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671644 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-config-data-custom\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671663 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rncq5\" (UniqueName: \"kubernetes.io/projected/84de7aef-2467-44a1-ae92-c22d87035c0d-kube-api-access-rncq5\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.671712 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-config-data-custom\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.678914 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-logs\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.679388 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-config-data\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.693872 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-combined-ca-bundle\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.699640 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-config-data-custom\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.741037 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-75574c7dff-p5nq5"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.773148 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-combined-ca-bundle\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.773184 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-config-data\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.773256 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-config-data-custom\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.773274 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rncq5\" (UniqueName: \"kubernetes.io/projected/84de7aef-2467-44a1-ae92-c22d87035c0d-kube-api-access-rncq5\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.773345 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84de7aef-2467-44a1-ae92-c22d87035c0d-logs\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.773778 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84de7aef-2467-44a1-ae92-c22d87035c0d-logs\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.778401 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mx8f\" (UniqueName: \"kubernetes.io/projected/68f7e592-7d70-4c02-a8c5-9b31ffc06c1c-kube-api-access-8mx8f\") pod \"barbican-keystone-listener-5645f9f854-tfsnf\" (UID: \"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c\") " pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.786161 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-4kkk9"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.797685 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-combined-ca-bundle\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.820520 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-ftr94"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.822051 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.827789 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-config-data\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.835641 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rncq5\" (UniqueName: \"kubernetes.io/projected/84de7aef-2467-44a1-ae92-c22d87035c0d-kube-api-access-rncq5\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.847319 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-ftr94"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.851721 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.853669 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84de7aef-2467-44a1-ae92-c22d87035c0d-config-data-custom\") pod \"barbican-worker-75574c7dff-p5nq5\" (UID: \"84de7aef-2467-44a1-ae92-c22d87035c0d\") " pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.882582 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-75574c7dff-p5nq5" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.966159 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5b8858bbdd-kp5tm"] Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.967613 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.974981 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.987580 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpk4\" (UniqueName: \"kubernetes.io/projected/cc84e061-4084-445a-b6ef-1e015ebd5007-kube-api-access-lmpk4\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.987679 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.987723 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.987757 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.987809 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-svc\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:01 crc kubenswrapper[4758]: I1004 11:08:01.987854 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-config\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.037258 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b8858bbdd-kp5tm"] Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.068546 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090510 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090570 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-config\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090595 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpk4\" (UniqueName: \"kubernetes.io/projected/cc84e061-4084-445a-b6ef-1e015ebd5007-kube-api-access-lmpk4\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090621 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-combined-ca-bundle\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090671 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090711 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr4nw\" (UniqueName: \"kubernetes.io/projected/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-kube-api-access-cr4nw\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090738 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090764 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090796 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-logs\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090820 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-svc\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.090852 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data-custom\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.091774 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-config\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.092585 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.092907 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.093190 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-svc\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.093459 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.155951 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpk4\" (UniqueName: \"kubernetes.io/projected/cc84e061-4084-445a-b6ef-1e015ebd5007-kube-api-access-lmpk4\") pod \"dnsmasq-dns-85ff748b95-ftr94\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.192037 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-logs\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.192454 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data-custom\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.192480 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.192506 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-combined-ca-bundle\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.192577 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr4nw\" (UniqueName: \"kubernetes.io/projected/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-kube-api-access-cr4nw\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.193864 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-logs\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.214203 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.215137 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-combined-ca-bundle\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.216513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerStarted","Data":"de3c559a9a7aa0f655c985fb7967608cca32c1fa3a9484c66df87e87ab0abb64"} Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.218707 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data-custom\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.230605 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.236884 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr4nw\" (UniqueName: \"kubernetes.io/projected/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-kube-api-access-cr4nw\") pod \"barbican-api-5b8858bbdd-kp5tm\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.259609 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" event={"ID":"5b91f490-3d53-4563-99ee-2f58cbfba4f2","Type":"ContainerStarted","Data":"ba599ce27b74c5fd1e2a5d2c3b1538bcbf1b6d81bb483b75c533abee6023a878"} Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.260483 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.286053 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.308758 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d66f67fd7-kd9f7" event={"ID":"f3b292da-54e8-4646-b70a-b2618eb4af4d","Type":"ContainerStarted","Data":"cceeb04efccb2786d61c5c6f96e28280d39fa5da6dd835242255fa75e86f318d"} Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.310352 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" podStartSLOduration=12.310342365 podStartE2EDuration="12.310342365s" podCreationTimestamp="2025-10-04 11:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:02.310175 +0000 UTC m=+1059.602825889" watchObservedRunningTime="2025-10-04 11:08:02.310342365 +0000 UTC m=+1059.602993254" Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.621730 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ccf54cd5-74s2f"] Oct 04 11:08:02 crc kubenswrapper[4758]: I1004 11:08:02.850700 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5645f9f854-tfsnf"] Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.322967 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccf54cd5-74s2f" event={"ID":"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce","Type":"ContainerStarted","Data":"eec4c05ebd492b7b77430a8ebdcdb06c0328d4d73fb7921bc1b4f85ec9d8508e"} Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.339762 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d66f67fd7-kd9f7" event={"ID":"f3b292da-54e8-4646-b70a-b2618eb4af4d","Type":"ContainerStarted","Data":"4971e8345c39bfa37d1e259b2ed6e35f1bf9cd153e8d39d36234b21d7579c61b"} Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.340765 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6ad8010-4207-4d04-9765-5b81facf9cad","Type":"ContainerStarted","Data":"4ab23623bbd35983978197a60e4cbe69801b418dee40c090e20920c56ad37e18"} Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.340870 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.341014 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-75574c7dff-p5nq5"] Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.355277 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d059c9e-fa03-46dd-b113-de161b74ecea","Type":"ContainerStarted","Data":"c505624a2c304042dbfb72ea933c5983844a1a26b39789dcdaf77e0eb5c59df2"} Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.357703 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerName="dnsmasq-dns" containerID="cri-o://ba599ce27b74c5fd1e2a5d2c3b1538bcbf1b6d81bb483b75c533abee6023a878" gracePeriod=10 Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.357974 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" event={"ID":"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c","Type":"ContainerStarted","Data":"7f476031c1aa9f5c5993dccd7e02f794beb321d4523ff732e028fcba02854649"} Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.417065 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-ftr94"] Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.442095 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8488664d88-ghdnd"] Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.460243 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b8858bbdd-kp5tm"] Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.479744 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.479726048 podStartE2EDuration="10.479726048s" podCreationTimestamp="2025-10-04 11:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:03.467343462 +0000 UTC m=+1060.759994491" watchObservedRunningTime="2025-10-04 11:08:03.479726048 +0000 UTC m=+1060.772376937" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.487485 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.487470648 podStartE2EDuration="11.487470648s" podCreationTimestamp="2025-10-04 11:07:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:03.486582644 +0000 UTC m=+1060.779233533" watchObservedRunningTime="2025-10-04 11:08:03.487470648 +0000 UTC m=+1060.780121537" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.530815 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d66f67fd7-kd9f7" podStartSLOduration=9.530799843 podStartE2EDuration="9.530799843s" podCreationTimestamp="2025-10-04 11:07:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:03.525620153 +0000 UTC m=+1060.818271042" watchObservedRunningTime="2025-10-04 11:08:03.530799843 +0000 UTC m=+1060.823450722" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.732705 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.734974 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.768327 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:03 crc kubenswrapper[4758]: I1004 11:08:03.787485 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:04 crc kubenswrapper[4758]: I1004 11:08:04.367251 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccf54cd5-74s2f" event={"ID":"6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce","Type":"ContainerStarted","Data":"6673f15a1a6259cfeeeb51cacfd764f1874c252cfff3e719f03221ec4a35f1ea"} Oct 04 11:08:04 crc kubenswrapper[4758]: I1004 11:08:04.369260 4758 generic.go:334] "Generic (PLEG): container finished" podID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerID="ba599ce27b74c5fd1e2a5d2c3b1538bcbf1b6d81bb483b75c533abee6023a878" exitCode=0 Oct 04 11:08:04 crc kubenswrapper[4758]: I1004 11:08:04.370438 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" event={"ID":"5b91f490-3d53-4563-99ee-2f58cbfba4f2","Type":"ContainerDied","Data":"ba599ce27b74c5fd1e2a5d2c3b1538bcbf1b6d81bb483b75c533abee6023a878"} Oct 04 11:08:04 crc kubenswrapper[4758]: I1004 11:08:04.370510 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:04 crc kubenswrapper[4758]: I1004 11:08:04.371525 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:04 crc kubenswrapper[4758]: W1004 11:08:04.514734 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96e4bc74_27b8_4c61_85c5_d56589b9e4bf.slice/crio-2e89e1c45445f5f49c92137d34bdc7ec54b0ce522492ee4c8a68eb3b764b08a2 WatchSource:0}: Error finding container 2e89e1c45445f5f49c92137d34bdc7ec54b0ce522492ee4c8a68eb3b764b08a2: Status 404 returned error can't find the container with id 2e89e1c45445f5f49c92137d34bdc7ec54b0ce522492ee4c8a68eb3b764b08a2 Oct 04 11:08:04 crc kubenswrapper[4758]: I1004 11:08:04.958444 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.114609 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-config\") pod \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.114967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-svc\") pod \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.115059 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8x77\" (UniqueName: \"kubernetes.io/projected/5b91f490-3d53-4563-99ee-2f58cbfba4f2-kube-api-access-s8x77\") pod \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.115231 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-swift-storage-0\") pod \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.115264 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-sb\") pod \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.115289 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-nb\") pod \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\" (UID: \"5b91f490-3d53-4563-99ee-2f58cbfba4f2\") " Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.144311 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b91f490-3d53-4563-99ee-2f58cbfba4f2-kube-api-access-s8x77" (OuterVolumeSpecName: "kube-api-access-s8x77") pod "5b91f490-3d53-4563-99ee-2f58cbfba4f2" (UID: "5b91f490-3d53-4563-99ee-2f58cbfba4f2"). InnerVolumeSpecName "kube-api-access-s8x77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.223193 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8x77\" (UniqueName: \"kubernetes.io/projected/5b91f490-3d53-4563-99ee-2f58cbfba4f2-kube-api-access-s8x77\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.270330 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5b91f490-3d53-4563-99ee-2f58cbfba4f2" (UID: "5b91f490-3d53-4563-99ee-2f58cbfba4f2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.306470 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-config" (OuterVolumeSpecName: "config") pod "5b91f490-3d53-4563-99ee-2f58cbfba4f2" (UID: "5b91f490-3d53-4563-99ee-2f58cbfba4f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.306762 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5b91f490-3d53-4563-99ee-2f58cbfba4f2" (UID: "5b91f490-3d53-4563-99ee-2f58cbfba4f2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.325404 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.325442 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.325454 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.346302 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5b91f490-3d53-4563-99ee-2f58cbfba4f2" (UID: "5b91f490-3d53-4563-99ee-2f58cbfba4f2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.417243 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5b91f490-3d53-4563-99ee-2f58cbfba4f2" (UID: "5b91f490-3d53-4563-99ee-2f58cbfba4f2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.418904 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c9d6c44f4-gx842"] Oct 04 11:08:05 crc kubenswrapper[4758]: E1004 11:08:05.419186 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerName="init" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.419197 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerName="init" Oct 04 11:08:05 crc kubenswrapper[4758]: E1004 11:08:05.419222 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerName="dnsmasq-dns" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.419229 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerName="dnsmasq-dns" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.419376 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" containerName="dnsmasq-dns" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.427278 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.427305 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5b91f490-3d53-4563-99ee-2f58cbfba4f2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.441645 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c9d6c44f4-gx842"] Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.441750 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.444471 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.444749 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.476347 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75574c7dff-p5nq5" event={"ID":"84de7aef-2467-44a1-ae92-c22d87035c0d","Type":"ContainerStarted","Data":"8611fd55095bc67491769519dfa9f7a005c03d92c50bdd6e3edb393b12b9426c"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.539370 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8858bbdd-kp5tm" event={"ID":"aaa3a018-52a6-423d-ba2b-37ef1d0fab71","Type":"ContainerStarted","Data":"d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.539416 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8858bbdd-kp5tm" event={"ID":"aaa3a018-52a6-423d-ba2b-37ef1d0fab71","Type":"ContainerStarted","Data":"a2067c96bdb8953934a0f67d182799be504fd3a75c34c596cdaa55037f6d6b02"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.540438 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84xmm\" (UniqueName: \"kubernetes.io/projected/4df5fcde-8552-4277-896b-e5f47e9b64b1-kube-api-access-84xmm\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.540724 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-internal-tls-certs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.540817 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-config-data\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.540918 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-combined-ca-bundle\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.541003 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4df5fcde-8552-4277-896b-e5f47e9b64b1-logs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.541083 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-public-tls-certs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.541176 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-config-data-custom\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.597996 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerID="45f4c106cee77e1069cacb715c40e9166ffe8cd2de6978027982ec139f81a11b" exitCode=0 Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.598074 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" event={"ID":"cc84e061-4084-445a-b6ef-1e015ebd5007","Type":"ContainerDied","Data":"45f4c106cee77e1069cacb715c40e9166ffe8cd2de6978027982ec139f81a11b"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.598119 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" event={"ID":"cc84e061-4084-445a-b6ef-1e015ebd5007","Type":"ContainerStarted","Data":"bad93bc421f7d3c5f49bcfe83220b933bd47e2162e67ceb8bf4c296bc9c9d0cf"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.606996 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" event={"ID":"5b91f490-3d53-4563-99ee-2f58cbfba4f2","Type":"ContainerDied","Data":"71bed0efca6593802fb361aeeb957b817ffa5cb3b5ea63758f959610c16b262e"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.607045 4758 scope.go:117] "RemoveContainer" containerID="ba599ce27b74c5fd1e2a5d2c3b1538bcbf1b6d81bb483b75c533abee6023a878" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.607183 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-4kkk9" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.617507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8488664d88-ghdnd" event={"ID":"96e4bc74-27b8-4c61-85c5-d56589b9e4bf","Type":"ContainerStarted","Data":"5d67e31dea0471fcaa37de81dd171e37a84ad60d66178ae1158acb85ba597dc7"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.617552 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8488664d88-ghdnd" event={"ID":"96e4bc74-27b8-4c61-85c5-d56589b9e4bf","Type":"ContainerStarted","Data":"2e89e1c45445f5f49c92137d34bdc7ec54b0ce522492ee4c8a68eb3b764b08a2"} Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.618083 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642436 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-internal-tls-certs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642481 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-config-data\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642556 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-combined-ca-bundle\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642598 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4df5fcde-8552-4277-896b-e5f47e9b64b1-logs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642621 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-config-data-custom\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642640 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-public-tls-certs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.642680 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84xmm\" (UniqueName: \"kubernetes.io/projected/4df5fcde-8552-4277-896b-e5f47e9b64b1-kube-api-access-84xmm\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.647578 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4df5fcde-8552-4277-896b-e5f47e9b64b1-logs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.663905 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-config-data-custom\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.676785 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-ccf54cd5-74s2f" podStartSLOduration=4.676762648 podStartE2EDuration="4.676762648s" podCreationTimestamp="2025-10-04 11:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:05.665241066 +0000 UTC m=+1062.957891955" watchObservedRunningTime="2025-10-04 11:08:05.676762648 +0000 UTC m=+1062.969413537" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.717979 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-public-tls-certs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.718276 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-4kkk9"] Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.719015 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-config-data\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.727772 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-internal-tls-certs\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.729253 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df5fcde-8552-4277-896b-e5f47e9b64b1-combined-ca-bundle\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.730575 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84xmm\" (UniqueName: \"kubernetes.io/projected/4df5fcde-8552-4277-896b-e5f47e9b64b1-kube-api-access-84xmm\") pod \"barbican-api-6c9d6c44f4-gx842\" (UID: \"4df5fcde-8552-4277-896b-e5f47e9b64b1\") " pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.737917 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-4kkk9"] Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.761735 4758 scope.go:117] "RemoveContainer" containerID="a1837ae4301df85bb22d793762f5489552eac3aed4743a567b0bb8dd78761a2c" Oct 04 11:08:05 crc kubenswrapper[4758]: I1004 11:08:05.795127 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.388970 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c9d6c44f4-gx842"] Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.639448 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w9lbx" event={"ID":"41d2a43a-542a-40c5-884f-e6540082151e","Type":"ContainerStarted","Data":"b95adcbdc5be222cb3cb312af727b67de1985c050a2a53a7c6d7552caccd835c"} Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.646067 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8858bbdd-kp5tm" event={"ID":"aaa3a018-52a6-423d-ba2b-37ef1d0fab71","Type":"ContainerStarted","Data":"ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3"} Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.646572 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.646602 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.649850 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" event={"ID":"cc84e061-4084-445a-b6ef-1e015ebd5007","Type":"ContainerStarted","Data":"288db352cd496aa2fe0003f561f25e8ad21a7116b4f0d98b495aa33bacfa3c5d"} Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.650466 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.674511 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-w9lbx" podStartSLOduration=3.887349809 podStartE2EDuration="52.674490155s" podCreationTimestamp="2025-10-04 11:07:14 +0000 UTC" firstStartedPulling="2025-10-04 11:07:15.752783643 +0000 UTC m=+1013.045434522" lastFinishedPulling="2025-10-04 11:08:04.539923979 +0000 UTC m=+1061.832574868" observedRunningTime="2025-10-04 11:08:06.655659404 +0000 UTC m=+1063.948310293" watchObservedRunningTime="2025-10-04 11:08:06.674490155 +0000 UTC m=+1063.967141044" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.689531 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8488664d88-ghdnd" event={"ID":"96e4bc74-27b8-4c61-85c5-d56589b9e4bf","Type":"ContainerStarted","Data":"c6a02c2d13b850d7b95a247bcf5bcadd43017c362f57b512891c82307b374bef"} Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.689566 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.689602 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.689999 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.695806 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" podStartSLOduration=5.695788203 podStartE2EDuration="5.695788203s" podCreationTimestamp="2025-10-04 11:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:06.671495794 +0000 UTC m=+1063.964146683" watchObservedRunningTime="2025-10-04 11:08:06.695788203 +0000 UTC m=+1063.988439092" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.711493 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5b8858bbdd-kp5tm" podStartSLOduration=5.711476678 podStartE2EDuration="5.711476678s" podCreationTimestamp="2025-10-04 11:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:06.697054817 +0000 UTC m=+1063.989705706" watchObservedRunningTime="2025-10-04 11:08:06.711476678 +0000 UTC m=+1064.004127567" Oct 04 11:08:06 crc kubenswrapper[4758]: I1004 11:08:06.727233 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8488664d88-ghdnd" podStartSLOduration=5.727218156 podStartE2EDuration="5.727218156s" podCreationTimestamp="2025-10-04 11:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:06.718241722 +0000 UTC m=+1064.010892611" watchObservedRunningTime="2025-10-04 11:08:06.727218156 +0000 UTC m=+1064.019869045" Oct 04 11:08:07 crc kubenswrapper[4758]: I1004 11:08:07.340347 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b91f490-3d53-4563-99ee-2f58cbfba4f2" path="/var/lib/kubelet/pods/5b91f490-3d53-4563-99ee-2f58cbfba4f2/volumes" Oct 04 11:08:07 crc kubenswrapper[4758]: I1004 11:08:07.700375 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9d6c44f4-gx842" event={"ID":"4df5fcde-8552-4277-896b-e5f47e9b64b1","Type":"ContainerStarted","Data":"df3ad7b5ae04ddbcbd582d36608db6f778c34494e7d5d2699ca63f6c51aaaff3"} Oct 04 11:08:08 crc kubenswrapper[4758]: I1004 11:08:08.578746 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 04 11:08:08 crc kubenswrapper[4758]: I1004 11:08:08.763767 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75574c7dff-p5nq5" event={"ID":"84de7aef-2467-44a1-ae92-c22d87035c0d","Type":"ContainerStarted","Data":"2ab0fbe2291834876ee225304c5d6191404c33f35b67f940c0e8d05152afe141"} Oct 04 11:08:08 crc kubenswrapper[4758]: I1004 11:08:08.765415 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c98cb656b-lczgq" podUID="61f721a8-00fb-446c-8e98-922827d2fc39" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 11:08:08 crc kubenswrapper[4758]: I1004 11:08:08.772084 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" event={"ID":"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c","Type":"ContainerStarted","Data":"cd77f77c09422a5b01e85538c6a76d0f6a4366846a5a81c10fe228d80db16bbb"} Oct 04 11:08:08 crc kubenswrapper[4758]: I1004 11:08:08.775048 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9d6c44f4-gx842" event={"ID":"4df5fcde-8552-4277-896b-e5f47e9b64b1","Type":"ContainerStarted","Data":"f3bdf56d4f02b8a4ed4f780632b8ca980850d7830c3c5871789cb52efe53e3af"} Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.809161 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" event={"ID":"68f7e592-7d70-4c02-a8c5-9b31ffc06c1c","Type":"ContainerStarted","Data":"3d069ff0db17aeccb675f484486e1cb6db7046f385fa82e000b9d8f4c77601f9"} Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.830092 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9d6c44f4-gx842" event={"ID":"4df5fcde-8552-4277-896b-e5f47e9b64b1","Type":"ContainerStarted","Data":"cb01f42f73ef93ecbed41855101011ca94852fcc77a8481a3b04ccc624343801"} Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.830174 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.830187 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.858573 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75574c7dff-p5nq5" event={"ID":"84de7aef-2467-44a1-ae92-c22d87035c0d","Type":"ContainerStarted","Data":"e5b437bf5e92096062546410dde0ebbd31c11b271060f9d63e051a0faa19735e"} Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.859164 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5645f9f854-tfsnf" podStartSLOduration=3.790499316 podStartE2EDuration="8.859152499s" podCreationTimestamp="2025-10-04 11:08:01 +0000 UTC" firstStartedPulling="2025-10-04 11:08:02.956337769 +0000 UTC m=+1060.248988658" lastFinishedPulling="2025-10-04 11:08:08.024990952 +0000 UTC m=+1065.317641841" observedRunningTime="2025-10-04 11:08:09.828432196 +0000 UTC m=+1067.121083085" watchObservedRunningTime="2025-10-04 11:08:09.859152499 +0000 UTC m=+1067.151803388" Oct 04 11:08:09 crc kubenswrapper[4758]: I1004 11:08:09.861401 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c9d6c44f4-gx842" podStartSLOduration=4.86139211 podStartE2EDuration="4.86139211s" podCreationTimestamp="2025-10-04 11:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:09.849643411 +0000 UTC m=+1067.142294300" watchObservedRunningTime="2025-10-04 11:08:09.86139211 +0000 UTC m=+1067.154042999" Oct 04 11:08:10 crc kubenswrapper[4758]: I1004 11:08:10.535034 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:10 crc kubenswrapper[4758]: I1004 11:08:10.572135 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-75574c7dff-p5nq5" podStartSLOduration=6.046308312 podStartE2EDuration="9.572118121s" podCreationTimestamp="2025-10-04 11:08:01 +0000 UTC" firstStartedPulling="2025-10-04 11:08:04.501673181 +0000 UTC m=+1061.794324070" lastFinishedPulling="2025-10-04 11:08:08.02748299 +0000 UTC m=+1065.320133879" observedRunningTime="2025-10-04 11:08:09.877436795 +0000 UTC m=+1067.170087694" watchObservedRunningTime="2025-10-04 11:08:10.572118121 +0000 UTC m=+1067.864769010" Oct 04 11:08:12 crc kubenswrapper[4758]: I1004 11:08:12.233236 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:12 crc kubenswrapper[4758]: I1004 11:08:12.311716 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4dncc"] Oct 04 11:08:12 crc kubenswrapper[4758]: I1004 11:08:12.312004 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-4dncc" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="dnsmasq-dns" containerID="cri-o://22036001acb4ea1abcfa323d47d4df958c65d491a024808769b6ad1ba44f5294" gracePeriod=10 Oct 04 11:08:12 crc kubenswrapper[4758]: I1004 11:08:12.891612 4758 generic.go:334] "Generic (PLEG): container finished" podID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerID="22036001acb4ea1abcfa323d47d4df958c65d491a024808769b6ad1ba44f5294" exitCode=0 Oct 04 11:08:12 crc kubenswrapper[4758]: I1004 11:08:12.891786 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4dncc" event={"ID":"9f9584b6-569d-4e7a-bfd1-df47048bd3aa","Type":"ContainerDied","Data":"22036001acb4ea1abcfa323d47d4df958c65d491a024808769b6ad1ba44f5294"} Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.248482 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.248524 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.300750 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.375675 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.835025 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.900691 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 11:08:13 crc kubenswrapper[4758]: I1004 11:08:13.900810 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 11:08:15 crc kubenswrapper[4758]: I1004 11:08:15.134280 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:15 crc kubenswrapper[4758]: I1004 11:08:15.167266 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-4dncc" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Oct 04 11:08:15 crc kubenswrapper[4758]: I1004 11:08:15.487251 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:17 crc kubenswrapper[4758]: I1004 11:08:17.959212 4758 generic.go:334] "Generic (PLEG): container finished" podID="41d2a43a-542a-40c5-884f-e6540082151e" containerID="b95adcbdc5be222cb3cb312af727b67de1985c050a2a53a7c6d7552caccd835c" exitCode=0 Oct 04 11:08:17 crc kubenswrapper[4758]: I1004 11:08:17.959454 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w9lbx" event={"ID":"41d2a43a-542a-40c5-884f-e6540082151e","Type":"ContainerDied","Data":"b95adcbdc5be222cb3cb312af727b67de1985c050a2a53a7c6d7552caccd835c"} Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.069290 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.574574 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.755839 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c9d6c44f4-gx842" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.756341 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c98cb656b-lczgq" podUID="61f721a8-00fb-446c-8e98-922827d2fc39" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.756481 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.757431 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"1879f1801937d4659da7d57131cbee347b49df5d138e811f68d0a8385f43bfe5"} pod="openstack/horizon-5c98cb656b-lczgq" containerMessage="Container horizon failed startup probe, will be restarted" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.757472 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5c98cb656b-lczgq" podUID="61f721a8-00fb-446c-8e98-922827d2fc39" containerName="horizon" containerID="cri-o://1879f1801937d4659da7d57131cbee347b49df5d138e811f68d0a8385f43bfe5" gracePeriod=30 Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.786969 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.787069 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.826158 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b8858bbdd-kp5tm"] Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.826403 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b8858bbdd-kp5tm" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api-log" containerID="cri-o://d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d" gracePeriod=30 Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.826834 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b8858bbdd-kp5tm" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api" containerID="cri-o://ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3" gracePeriod=30 Oct 04 11:08:18 crc kubenswrapper[4758]: I1004 11:08:18.848613 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8858bbdd-kp5tm" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Oct 04 11:08:19 crc kubenswrapper[4758]: I1004 11:08:19.037457 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 11:08:19 crc kubenswrapper[4758]: I1004 11:08:19.987698 4758 generic.go:334] "Generic (PLEG): container finished" podID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerID="d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d" exitCode=143 Oct 04 11:08:19 crc kubenswrapper[4758]: I1004 11:08:19.987913 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8858bbdd-kp5tm" event={"ID":"aaa3a018-52a6-423d-ba2b-37ef1d0fab71","Type":"ContainerDied","Data":"d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d"} Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.165976 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-4dncc" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Oct 04 11:08:20 crc kubenswrapper[4758]: E1004 11:08:20.437661 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 04 11:08:20 crc kubenswrapper[4758]: E1004 11:08:20.438155 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pt49z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(d105f7c6-5dbf-42a0-9821-b1a31facaf5e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.439738 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:08:20 crc kubenswrapper[4758]: E1004 11:08:20.439816 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.602157 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-db-sync-config-data\") pod \"41d2a43a-542a-40c5-884f-e6540082151e\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.602198 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d2a43a-542a-40c5-884f-e6540082151e-etc-machine-id\") pod \"41d2a43a-542a-40c5-884f-e6540082151e\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.602236 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-scripts\") pod \"41d2a43a-542a-40c5-884f-e6540082151e\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.602316 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-combined-ca-bundle\") pod \"41d2a43a-542a-40c5-884f-e6540082151e\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.602398 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-config-data\") pod \"41d2a43a-542a-40c5-884f-e6540082151e\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.602468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r8j7\" (UniqueName: \"kubernetes.io/projected/41d2a43a-542a-40c5-884f-e6540082151e-kube-api-access-7r8j7\") pod \"41d2a43a-542a-40c5-884f-e6540082151e\" (UID: \"41d2a43a-542a-40c5-884f-e6540082151e\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.605308 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41d2a43a-542a-40c5-884f-e6540082151e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "41d2a43a-542a-40c5-884f-e6540082151e" (UID: "41d2a43a-542a-40c5-884f-e6540082151e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.611782 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-scripts" (OuterVolumeSpecName: "scripts") pod "41d2a43a-542a-40c5-884f-e6540082151e" (UID: "41d2a43a-542a-40c5-884f-e6540082151e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.612284 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d2a43a-542a-40c5-884f-e6540082151e-kube-api-access-7r8j7" (OuterVolumeSpecName: "kube-api-access-7r8j7") pod "41d2a43a-542a-40c5-884f-e6540082151e" (UID: "41d2a43a-542a-40c5-884f-e6540082151e"). InnerVolumeSpecName "kube-api-access-7r8j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.615188 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "41d2a43a-542a-40c5-884f-e6540082151e" (UID: "41d2a43a-542a-40c5-884f-e6540082151e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.664336 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41d2a43a-542a-40c5-884f-e6540082151e" (UID: "41d2a43a-542a-40c5-884f-e6540082151e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.673742 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-config-data" (OuterVolumeSpecName: "config-data") pod "41d2a43a-542a-40c5-884f-e6540082151e" (UID: "41d2a43a-542a-40c5-884f-e6540082151e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.704448 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.704483 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.704493 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r8j7\" (UniqueName: \"kubernetes.io/projected/41d2a43a-542a-40c5-884f-e6540082151e-kube-api-access-7r8j7\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.704505 4758 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.704514 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d2a43a-542a-40c5-884f-e6540082151e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.704522 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d2a43a-542a-40c5-884f-e6540082151e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.745114 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.908112 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-nb\") pod \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.908165 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-dns-svc\") pod \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.908242 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc758\" (UniqueName: \"kubernetes.io/projected/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-kube-api-access-tc758\") pod \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.908304 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-sb\") pod \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.908346 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-config\") pod \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\" (UID: \"9f9584b6-569d-4e7a-bfd1-df47048bd3aa\") " Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.921285 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-kube-api-access-tc758" (OuterVolumeSpecName: "kube-api-access-tc758") pod "9f9584b6-569d-4e7a-bfd1-df47048bd3aa" (UID: "9f9584b6-569d-4e7a-bfd1-df47048bd3aa"). InnerVolumeSpecName "kube-api-access-tc758". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.968785 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f9584b6-569d-4e7a-bfd1-df47048bd3aa" (UID: "9f9584b6-569d-4e7a-bfd1-df47048bd3aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.976257 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f9584b6-569d-4e7a-bfd1-df47048bd3aa" (UID: "9f9584b6-569d-4e7a-bfd1-df47048bd3aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.978172 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f9584b6-569d-4e7a-bfd1-df47048bd3aa" (UID: "9f9584b6-569d-4e7a-bfd1-df47048bd3aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.994176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-config" (OuterVolumeSpecName: "config") pod "9f9584b6-569d-4e7a-bfd1-df47048bd3aa" (UID: "9f9584b6-569d-4e7a-bfd1-df47048bd3aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.998241 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-w9lbx" event={"ID":"41d2a43a-542a-40c5-884f-e6540082151e","Type":"ContainerDied","Data":"5d92bdb70f436860726f1bae62926df758cfd39693368881eec18cf68d3ff6d4"} Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.998280 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d92bdb70f436860726f1bae62926df758cfd39693368881eec18cf68d3ff6d4" Oct 04 11:08:20 crc kubenswrapper[4758]: I1004 11:08:20.998329 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-w9lbx" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.001510 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-central-agent" containerID="cri-o://d568eaf29b4ed2db926b7cb81837f209219d5272453d804aec6a62e9c7db4624" gracePeriod=30 Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.001616 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4dncc" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.003216 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4dncc" event={"ID":"9f9584b6-569d-4e7a-bfd1-df47048bd3aa","Type":"ContainerDied","Data":"c28fd65c5dafcc86ade061d0837115c8e24e9f69b521df9d168640a598016c43"} Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.003272 4758 scope.go:117] "RemoveContainer" containerID="22036001acb4ea1abcfa323d47d4df958c65d491a024808769b6ad1ba44f5294" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.003753 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="sg-core" containerID="cri-o://de3c559a9a7aa0f655c985fb7967608cca32c1fa3a9484c66df87e87ab0abb64" gracePeriod=30 Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.003849 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-notification-agent" containerID="cri-o://d729e35291e40a7db33b5296b8d19ad1aa2eadda21d1a98483af40a691588f69" gracePeriod=30 Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.013085 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.013133 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.013143 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc758\" (UniqueName: \"kubernetes.io/projected/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-kube-api-access-tc758\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.013155 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.013176 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f9584b6-569d-4e7a-bfd1-df47048bd3aa-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.023796 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.053746 4758 scope.go:117] "RemoveContainer" containerID="88e13b31bc6f1c35db9adbbd675c0e2c908a20cf19cce67277c3bdd1275f059b" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.075021 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4dncc"] Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.086479 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4dncc"] Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.336240 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" path="/var/lib/kubelet/pods/9f9584b6-569d-4e7a-bfd1-df47048bd3aa/volumes" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.728280 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:21 crc kubenswrapper[4758]: E1004 11:08:21.728941 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d2a43a-542a-40c5-884f-e6540082151e" containerName="cinder-db-sync" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.728959 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d2a43a-542a-40c5-884f-e6540082151e" containerName="cinder-db-sync" Oct 04 11:08:21 crc kubenswrapper[4758]: E1004 11:08:21.728971 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="init" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.728977 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="init" Oct 04 11:08:21 crc kubenswrapper[4758]: E1004 11:08:21.729002 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="dnsmasq-dns" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.729009 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="dnsmasq-dns" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.729190 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d2a43a-542a-40c5-884f-e6540082151e" containerName="cinder-db-sync" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.729202 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f9584b6-569d-4e7a-bfd1-df47048bd3aa" containerName="dnsmasq-dns" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.730065 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.737917 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.740495 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.740640 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-np5w8" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.744080 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.757957 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.824966 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.825025 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkffm\" (UniqueName: \"kubernetes.io/projected/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-kube-api-access-xkffm\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.825075 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.825280 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.825354 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.825439 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.866243 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kt56w"] Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.867997 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.926764 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.926827 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkffm\" (UniqueName: \"kubernetes.io/projected/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-kube-api-access-xkffm\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.926864 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.926882 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.927222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.927251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.927283 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.933738 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.939570 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-scripts\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.939881 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.956350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:21 crc kubenswrapper[4758]: I1004 11:08:21.989188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkffm\" (UniqueName: \"kubernetes.io/projected/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-kube-api-access-xkffm\") pod \"cinder-scheduler-0\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.015829 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kt56w"] Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.028849 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-config\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.028913 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.028938 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.028960 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.028999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz4rc\" (UniqueName: \"kubernetes.io/projected/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-kube-api-access-qz4rc\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.029019 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.047731 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.078417 4758 generic.go:334] "Generic (PLEG): container finished" podID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerID="de3c559a9a7aa0f655c985fb7967608cca32c1fa3a9484c66df87e87ab0abb64" exitCode=2 Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.078639 4758 generic.go:334] "Generic (PLEG): container finished" podID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerID="d568eaf29b4ed2db926b7cb81837f209219d5272453d804aec6a62e9c7db4624" exitCode=0 Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.078751 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerDied","Data":"de3c559a9a7aa0f655c985fb7967608cca32c1fa3a9484c66df87e87ab0abb64"} Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.078842 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerDied","Data":"d568eaf29b4ed2db926b7cb81837f209219d5272453d804aec6a62e9c7db4624"} Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.131045 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-config\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.131121 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.131144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.131161 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.131201 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz4rc\" (UniqueName: \"kubernetes.io/projected/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-kube-api-access-qz4rc\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.131217 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.132188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.132667 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-config\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.133222 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.133714 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.134222 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.154265 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz4rc\" (UniqueName: \"kubernetes.io/projected/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-kube-api-access-qz4rc\") pod \"dnsmasq-dns-5c9776ccc5-kt56w\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.180457 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.181953 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.187627 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.203740 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.221204 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337491 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb950bf-6e02-47a1-8134-0d4c36c90f43-logs\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337566 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data-custom\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337596 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337627 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bb950bf-6e02-47a1-8134-0d4c36c90f43-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337683 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stk48\" (UniqueName: \"kubernetes.io/projected/1bb950bf-6e02-47a1-8134-0d4c36c90f43-kube-api-access-stk48\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337726 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.337762 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-scripts\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442029 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442440 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-scripts\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442469 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb950bf-6e02-47a1-8134-0d4c36c90f43-logs\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442520 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data-custom\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442582 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bb950bf-6e02-47a1-8134-0d4c36c90f43-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.442629 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stk48\" (UniqueName: \"kubernetes.io/projected/1bb950bf-6e02-47a1-8134-0d4c36c90f43-kube-api-access-stk48\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.443533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb950bf-6e02-47a1-8134-0d4c36c90f43-logs\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.443563 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bb950bf-6e02-47a1-8134-0d4c36c90f43-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.461470 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.461735 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-scripts\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.465232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stk48\" (UniqueName: \"kubernetes.io/projected/1bb950bf-6e02-47a1-8134-0d4c36c90f43-kube-api-access-stk48\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.465409 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data-custom\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.471631 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.522525 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.814295 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kt56w"] Oct 04 11:08:22 crc kubenswrapper[4758]: I1004 11:08:22.938295 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:23 crc kubenswrapper[4758]: I1004 11:08:23.103848 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:23 crc kubenswrapper[4758]: I1004 11:08:23.110712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" event={"ID":"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7","Type":"ContainerStarted","Data":"b8a3bbe4a74470d6a1ec5495324ee0d2e9f0154bcfe295b824746a17c1a4cb67"} Oct 04 11:08:23 crc kubenswrapper[4758]: I1004 11:08:23.118806 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6","Type":"ContainerStarted","Data":"0afa976774a895c7fe6880049cee38fda4b23bbb494259cabd02405b220d4ec5"} Oct 04 11:08:23 crc kubenswrapper[4758]: I1004 11:08:23.423754 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8858bbdd-kp5tm" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:48614->10.217.0.158:9311: read: connection reset by peer" Oct 04 11:08:23 crc kubenswrapper[4758]: I1004 11:08:23.423761 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8858bbdd-kp5tm" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:48624->10.217.0.158:9311: read: connection reset by peer" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.109299 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.138221 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bb950bf-6e02-47a1-8134-0d4c36c90f43","Type":"ContainerStarted","Data":"a1752fe72b843186dc015152651802d7fd2c7d78819ee31dd02e1253f0efc8a4"} Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.142499 4758 generic.go:334] "Generic (PLEG): container finished" podID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerID="63753f506ade0576037bc099f5bf3b701c1f6e79c6247ab651a4250d1f075b36" exitCode=0 Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.142590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" event={"ID":"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7","Type":"ContainerDied","Data":"63753f506ade0576037bc099f5bf3b701c1f6e79c6247ab651a4250d1f075b36"} Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.170399 4758 generic.go:334] "Generic (PLEG): container finished" podID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerID="ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3" exitCode=0 Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.170440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8858bbdd-kp5tm" event={"ID":"aaa3a018-52a6-423d-ba2b-37ef1d0fab71","Type":"ContainerDied","Data":"ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3"} Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.170467 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8858bbdd-kp5tm" event={"ID":"aaa3a018-52a6-423d-ba2b-37ef1d0fab71","Type":"ContainerDied","Data":"a2067c96bdb8953934a0f67d182799be504fd3a75c34c596cdaa55037f6d6b02"} Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.170668 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8858bbdd-kp5tm" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.172577 4758 scope.go:117] "RemoveContainer" containerID="ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.199748 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-logs\") pod \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.199824 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr4nw\" (UniqueName: \"kubernetes.io/projected/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-kube-api-access-cr4nw\") pod \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.199854 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data-custom\") pod \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.199894 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-combined-ca-bundle\") pod \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.199993 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data\") pod \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\" (UID: \"aaa3a018-52a6-423d-ba2b-37ef1d0fab71\") " Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.201281 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-logs" (OuterVolumeSpecName: "logs") pod "aaa3a018-52a6-423d-ba2b-37ef1d0fab71" (UID: "aaa3a018-52a6-423d-ba2b-37ef1d0fab71"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.231432 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-kube-api-access-cr4nw" (OuterVolumeSpecName: "kube-api-access-cr4nw") pod "aaa3a018-52a6-423d-ba2b-37ef1d0fab71" (UID: "aaa3a018-52a6-423d-ba2b-37ef1d0fab71"). InnerVolumeSpecName "kube-api-access-cr4nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.250950 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aaa3a018-52a6-423d-ba2b-37ef1d0fab71" (UID: "aaa3a018-52a6-423d-ba2b-37ef1d0fab71"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.263469 4758 scope.go:117] "RemoveContainer" containerID="d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.268620 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaa3a018-52a6-423d-ba2b-37ef1d0fab71" (UID: "aaa3a018-52a6-423d-ba2b-37ef1d0fab71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.287406 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data" (OuterVolumeSpecName: "config-data") pod "aaa3a018-52a6-423d-ba2b-37ef1d0fab71" (UID: "aaa3a018-52a6-423d-ba2b-37ef1d0fab71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.301073 4758 scope.go:117] "RemoveContainer" containerID="ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.301948 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.301974 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.301989 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.302003 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr4nw\" (UniqueName: \"kubernetes.io/projected/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-kube-api-access-cr4nw\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.302015 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aaa3a018-52a6-423d-ba2b-37ef1d0fab71-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:24 crc kubenswrapper[4758]: E1004 11:08:24.309609 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3\": container with ID starting with ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3 not found: ID does not exist" containerID="ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.309658 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3"} err="failed to get container status \"ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3\": rpc error: code = NotFound desc = could not find container \"ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3\": container with ID starting with ad853d8f83925de6781b812603fddf0978e1edf556a0f87968beaec8662a9aa3 not found: ID does not exist" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.309693 4758 scope.go:117] "RemoveContainer" containerID="d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d" Oct 04 11:08:24 crc kubenswrapper[4758]: E1004 11:08:24.310365 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d\": container with ID starting with d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d not found: ID does not exist" containerID="d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.310390 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d"} err="failed to get container status \"d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d\": rpc error: code = NotFound desc = could not find container \"d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d\": container with ID starting with d7971dd6f0fafb4073d6ff340973c5359f5384edc053b8d63902b0521c57097d not found: ID does not exist" Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.523795 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b8858bbdd-kp5tm"] Oct 04 11:08:24 crc kubenswrapper[4758]: I1004 11:08:24.578335 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5b8858bbdd-kp5tm"] Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.126768 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7d66f67fd7-kd9f7" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.215585 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66cf6c6588-6rzc8"] Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.215791 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66cf6c6588-6rzc8" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-api" containerID="cri-o://7a57b136c7ec5a8825880144eddb8123052bc39b0a8532fc339e04bf6f56bd96" gracePeriod=30 Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.216189 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66cf6c6588-6rzc8" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-httpd" containerID="cri-o://b9ce32a8fce8271bdc3fbe67f0eee4af8f3430da4115e44584099b07cd3170c3" gracePeriod=30 Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.294664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6","Type":"ContainerStarted","Data":"a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f"} Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.305176 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bb950bf-6e02-47a1-8134-0d4c36c90f43","Type":"ContainerStarted","Data":"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4"} Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.314492 4758 generic.go:334] "Generic (PLEG): container finished" podID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerID="d729e35291e40a7db33b5296b8d19ad1aa2eadda21d1a98483af40a691588f69" exitCode=0 Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.314585 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerDied","Data":"d729e35291e40a7db33b5296b8d19ad1aa2eadda21d1a98483af40a691588f69"} Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.317319 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" event={"ID":"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7","Type":"ContainerStarted","Data":"eb11b7a9820a6929a2648090834f93daf33a7376779b897f063a5b27b3c74b93"} Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.318494 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.358080 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" path="/var/lib/kubelet/pods/aaa3a018-52a6-423d-ba2b-37ef1d0fab71/volumes" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.387485 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" podStartSLOduration=4.3874626039999995 podStartE2EDuration="4.387462604s" podCreationTimestamp="2025-10-04 11:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:25.360535594 +0000 UTC m=+1082.653186483" watchObservedRunningTime="2025-10-04 11:08:25.387462604 +0000 UTC m=+1082.680113483" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.787387 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.796863 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881034 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-log-httpd\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881343 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-config-data\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881460 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-run-httpd\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881577 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt49z\" (UniqueName: \"kubernetes.io/projected/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-kube-api-access-pt49z\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881677 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-combined-ca-bundle\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881759 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-sg-core-conf-yaml\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881861 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-scripts\") pod \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\" (UID: \"d105f7c6-5dbf-42a0-9821-b1a31facaf5e\") " Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.881473 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.882789 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.891484 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.908254 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-kube-api-access-pt49z" (OuterVolumeSpecName: "kube-api-access-pt49z") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "kube-api-access-pt49z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.923252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-scripts" (OuterVolumeSpecName: "scripts") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.970248 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.985022 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.985292 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt49z\" (UniqueName: \"kubernetes.io/projected/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-kube-api-access-pt49z\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.985369 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:25 crc kubenswrapper[4758]: I1004 11:08:25.985446 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.025946 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-config-data" (OuterVolumeSpecName: "config-data") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.032230 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d105f7c6-5dbf-42a0-9821-b1a31facaf5e" (UID: "d105f7c6-5dbf-42a0-9821-b1a31facaf5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.087393 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.087427 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d105f7c6-5dbf-42a0-9821-b1a31facaf5e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.341534 4758 generic.go:334] "Generic (PLEG): container finished" podID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerID="b9ce32a8fce8271bdc3fbe67f0eee4af8f3430da4115e44584099b07cd3170c3" exitCode=0 Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.341588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cf6c6588-6rzc8" event={"ID":"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb","Type":"ContainerDied","Data":"b9ce32a8fce8271bdc3fbe67f0eee4af8f3430da4115e44584099b07cd3170c3"} Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.343643 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6","Type":"ContainerStarted","Data":"c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000"} Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.345911 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bb950bf-6e02-47a1-8134-0d4c36c90f43","Type":"ContainerStarted","Data":"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c"} Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.346239 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.348633 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.353712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d105f7c6-5dbf-42a0-9821-b1a31facaf5e","Type":"ContainerDied","Data":"39b0eecef4325d0dd233eabe5622d7d5f3f0d1ca4d5b1cb933f9a6bce8923d07"} Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.353748 4758 scope.go:117] "RemoveContainer" containerID="de3c559a9a7aa0f655c985fb7967608cca32c1fa3a9484c66df87e87ab0abb64" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.377689 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.358189809 podStartE2EDuration="5.377672666s" podCreationTimestamp="2025-10-04 11:08:21 +0000 UTC" firstStartedPulling="2025-10-04 11:08:22.968331547 +0000 UTC m=+1080.260982436" lastFinishedPulling="2025-10-04 11:08:23.987814404 +0000 UTC m=+1081.280465293" observedRunningTime="2025-10-04 11:08:26.371158199 +0000 UTC m=+1083.663809088" watchObservedRunningTime="2025-10-04 11:08:26.377672666 +0000 UTC m=+1083.670323545" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.393285 4758 scope.go:117] "RemoveContainer" containerID="d729e35291e40a7db33b5296b8d19ad1aa2eadda21d1a98483af40a691588f69" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.400491 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.400458994 podStartE2EDuration="4.400458994s" podCreationTimestamp="2025-10-04 11:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:26.39587423 +0000 UTC m=+1083.688525119" watchObservedRunningTime="2025-10-04 11:08:26.400458994 +0000 UTC m=+1083.693109883" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.443467 4758 scope.go:117] "RemoveContainer" containerID="d568eaf29b4ed2db926b7cb81837f209219d5272453d804aec6a62e9c7db4624" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.465065 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.465123 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.485876 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:26 crc kubenswrapper[4758]: E1004 11:08:26.486297 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="sg-core" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486310 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="sg-core" Oct 04 11:08:26 crc kubenswrapper[4758]: E1004 11:08:26.486324 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486330 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api" Oct 04 11:08:26 crc kubenswrapper[4758]: E1004 11:08:26.486340 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-notification-agent" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486346 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-notification-agent" Oct 04 11:08:26 crc kubenswrapper[4758]: E1004 11:08:26.486359 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api-log" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486364 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api-log" Oct 04 11:08:26 crc kubenswrapper[4758]: E1004 11:08:26.486383 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-central-agent" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486388 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-central-agent" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486553 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="sg-core" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486565 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486580 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-notification-agent" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486593 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" containerName="ceilometer-central-agent" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.486606 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa3a018-52a6-423d-ba2b-37ef1d0fab71" containerName="barbican-api-log" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.488087 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.494898 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.497792 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.541164 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598402 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj7r6\" (UniqueName: \"kubernetes.io/projected/517aff26-ac5d-4395-9d0c-0d14690b7684-kube-api-access-nj7r6\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598761 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-log-httpd\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598804 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598881 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-run-httpd\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598915 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-scripts\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598956 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.598986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-config-data\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700210 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj7r6\" (UniqueName: \"kubernetes.io/projected/517aff26-ac5d-4395-9d0c-0d14690b7684-kube-api-access-nj7r6\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700267 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-log-httpd\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700295 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700377 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-run-httpd\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700394 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-scripts\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-config-data\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.700801 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-run-httpd\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.701021 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-log-httpd\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.710806 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-scripts\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.711217 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.715481 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.723745 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj7r6\" (UniqueName: \"kubernetes.io/projected/517aff26-ac5d-4395-9d0c-0d14690b7684-kube-api-access-nj7r6\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.724039 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-config-data\") pod \"ceilometer-0\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " pod="openstack/ceilometer-0" Oct 04 11:08:26 crc kubenswrapper[4758]: I1004 11:08:26.830993 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:27 crc kubenswrapper[4758]: I1004 11:08:27.050224 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 11:08:27 crc kubenswrapper[4758]: I1004 11:08:27.322297 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:27 crc kubenswrapper[4758]: W1004 11:08:27.332299 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod517aff26_ac5d_4395_9d0c_0d14690b7684.slice/crio-e714edbfd8b803163a5d1d981dcf115bdeccdefd2956befb8406e0b16ba8bee0 WatchSource:0}: Error finding container e714edbfd8b803163a5d1d981dcf115bdeccdefd2956befb8406e0b16ba8bee0: Status 404 returned error can't find the container with id e714edbfd8b803163a5d1d981dcf115bdeccdefd2956befb8406e0b16ba8bee0 Oct 04 11:08:27 crc kubenswrapper[4758]: I1004 11:08:27.381242 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d105f7c6-5dbf-42a0-9821-b1a31facaf5e" path="/var/lib/kubelet/pods/d105f7c6-5dbf-42a0-9821-b1a31facaf5e/volumes" Oct 04 11:08:27 crc kubenswrapper[4758]: I1004 11:08:27.387005 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerStarted","Data":"e714edbfd8b803163a5d1d981dcf115bdeccdefd2956befb8406e0b16ba8bee0"} Oct 04 11:08:27 crc kubenswrapper[4758]: I1004 11:08:27.389436 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api-log" containerID="cri-o://a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4" gracePeriod=30 Oct 04 11:08:27 crc kubenswrapper[4758]: I1004 11:08:27.390220 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api" containerID="cri-o://41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c" gracePeriod=30 Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.197270 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.355639 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-combined-ca-bundle\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.355895 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-scripts\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.355926 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bb950bf-6e02-47a1-8134-0d4c36c90f43-etc-machine-id\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.356062 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.356119 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data-custom\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.356186 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb950bf-6e02-47a1-8134-0d4c36c90f43-logs\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.356211 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stk48\" (UniqueName: \"kubernetes.io/projected/1bb950bf-6e02-47a1-8134-0d4c36c90f43-kube-api-access-stk48\") pod \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\" (UID: \"1bb950bf-6e02-47a1-8134-0d4c36c90f43\") " Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.359190 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1bb950bf-6e02-47a1-8134-0d4c36c90f43-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.362825 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb950bf-6e02-47a1-8134-0d4c36c90f43-kube-api-access-stk48" (OuterVolumeSpecName: "kube-api-access-stk48") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "kube-api-access-stk48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.366088 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bb950bf-6e02-47a1-8134-0d4c36c90f43-logs" (OuterVolumeSpecName: "logs") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.367176 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-scripts" (OuterVolumeSpecName: "scripts") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.383792 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.419245 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.419923 4758 generic.go:334] "Generic (PLEG): container finished" podID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerID="41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c" exitCode=0 Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.419951 4758 generic.go:334] "Generic (PLEG): container finished" podID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerID="a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4" exitCode=143 Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.420000 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bb950bf-6e02-47a1-8134-0d4c36c90f43","Type":"ContainerDied","Data":"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c"} Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.420027 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bb950bf-6e02-47a1-8134-0d4c36c90f43","Type":"ContainerDied","Data":"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4"} Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.420037 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1bb950bf-6e02-47a1-8134-0d4c36c90f43","Type":"ContainerDied","Data":"a1752fe72b843186dc015152651802d7fd2c7d78819ee31dd02e1253f0efc8a4"} Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.420051 4758 scope.go:117] "RemoveContainer" containerID="41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.420177 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.429959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerStarted","Data":"bd4f234257fdd115195d774554c71ecbc04fc9d271dbcd85d8680acbf29be583"} Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.430361 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data" (OuterVolumeSpecName: "config-data") pod "1bb950bf-6e02-47a1-8134-0d4c36c90f43" (UID: "1bb950bf-6e02-47a1-8134-0d4c36c90f43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.458940 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.458962 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.458974 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb950bf-6e02-47a1-8134-0d4c36c90f43-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.458983 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stk48\" (UniqueName: \"kubernetes.io/projected/1bb950bf-6e02-47a1-8134-0d4c36c90f43-kube-api-access-stk48\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.458991 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.458999 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bb950bf-6e02-47a1-8134-0d4c36c90f43-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.459008 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1bb950bf-6e02-47a1-8134-0d4c36c90f43-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.521070 4758 scope.go:117] "RemoveContainer" containerID="a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.538998 4758 scope.go:117] "RemoveContainer" containerID="41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c" Oct 04 11:08:28 crc kubenswrapper[4758]: E1004 11:08:28.539420 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c\": container with ID starting with 41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c not found: ID does not exist" containerID="41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.539476 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c"} err="failed to get container status \"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c\": rpc error: code = NotFound desc = could not find container \"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c\": container with ID starting with 41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c not found: ID does not exist" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.539506 4758 scope.go:117] "RemoveContainer" containerID="a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4" Oct 04 11:08:28 crc kubenswrapper[4758]: E1004 11:08:28.539758 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4\": container with ID starting with a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4 not found: ID does not exist" containerID="a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.539794 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4"} err="failed to get container status \"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4\": rpc error: code = NotFound desc = could not find container \"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4\": container with ID starting with a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4 not found: ID does not exist" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.539823 4758 scope.go:117] "RemoveContainer" containerID="41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.540058 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c"} err="failed to get container status \"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c\": rpc error: code = NotFound desc = could not find container \"41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c\": container with ID starting with 41d02c08d6839c89456260aa8736e766a1098792080b9383d7a090280b080f8c not found: ID does not exist" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.540083 4758 scope.go:117] "RemoveContainer" containerID="a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.540977 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4"} err="failed to get container status \"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4\": rpc error: code = NotFound desc = could not find container \"a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4\": container with ID starting with a98e96112567e21c73e08ad8c13d356b7e5d7912988329fcb4c4260fb11201a4 not found: ID does not exist" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.779952 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.790277 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.826165 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:28 crc kubenswrapper[4758]: E1004 11:08:28.826608 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.826627 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api" Oct 04 11:08:28 crc kubenswrapper[4758]: E1004 11:08:28.826652 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api-log" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.826660 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api-log" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.826844 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.826861 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" containerName="cinder-api-log" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.827813 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.830129 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.831223 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.835226 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.838488 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.974818 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-scripts\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.974854 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-config-data\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.974898 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-public-tls-certs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.975166 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sshx\" (UniqueName: \"kubernetes.io/projected/49fbaf4c-21ec-4246-86b8-1361ae5f3149-kube-api-access-2sshx\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.975222 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49fbaf4c-21ec-4246-86b8-1361ae5f3149-etc-machine-id\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.975318 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49fbaf4c-21ec-4246-86b8-1361ae5f3149-logs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.975420 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-config-data-custom\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.975478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:28 crc kubenswrapper[4758]: I1004 11:08:28.975524 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077055 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-config-data-custom\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077135 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077164 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077195 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-scripts\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077217 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-config-data\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077255 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-public-tls-certs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077312 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sshx\" (UniqueName: \"kubernetes.io/projected/49fbaf4c-21ec-4246-86b8-1361ae5f3149-kube-api-access-2sshx\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077341 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49fbaf4c-21ec-4246-86b8-1361ae5f3149-etc-machine-id\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49fbaf4c-21ec-4246-86b8-1361ae5f3149-logs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.077854 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49fbaf4c-21ec-4246-86b8-1361ae5f3149-logs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.083176 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49fbaf4c-21ec-4246-86b8-1361ae5f3149-etc-machine-id\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.083505 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.083982 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-config-data\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.085405 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-public-tls-certs\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.085500 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-scripts\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.086635 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-config-data-custom\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.087166 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49fbaf4c-21ec-4246-86b8-1361ae5f3149-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.096750 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sshx\" (UniqueName: \"kubernetes.io/projected/49fbaf4c-21ec-4246-86b8-1361ae5f3149-kube-api-access-2sshx\") pod \"cinder-api-0\" (UID: \"49fbaf4c-21ec-4246-86b8-1361ae5f3149\") " pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.161332 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.358122 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bb950bf-6e02-47a1-8134-0d4c36c90f43" path="/var/lib/kubelet/pods/1bb950bf-6e02-47a1-8134-0d4c36c90f43/volumes" Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.442555 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerStarted","Data":"5295b93ea52879cc60ec75d6f7a945518e30b60d3f4f500516f5cb8ddc5f7e91"} Oct 04 11:08:29 crc kubenswrapper[4758]: I1004 11:08:29.638934 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.480658 4758 generic.go:334] "Generic (PLEG): container finished" podID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerID="7a57b136c7ec5a8825880144eddb8123052bc39b0a8532fc339e04bf6f56bd96" exitCode=0 Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.480740 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cf6c6588-6rzc8" event={"ID":"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb","Type":"ContainerDied","Data":"7a57b136c7ec5a8825880144eddb8123052bc39b0a8532fc339e04bf6f56bd96"} Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.485081 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"49fbaf4c-21ec-4246-86b8-1361ae5f3149","Type":"ContainerStarted","Data":"7132a615eb8a2391f0ad4c47497495ac411992512a5e03edb072ce074a65ecec"} Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.485165 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"49fbaf4c-21ec-4246-86b8-1361ae5f3149","Type":"ContainerStarted","Data":"aca148498d04bd934225a2f52c24feb9309d1fb249bd3c60004052a50ae8e52b"} Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.487157 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerStarted","Data":"f9186911306d1a73e24f8f65029c47e948fc6f1de39edc8dcb9c0e9e53f5f473"} Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.628901 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.710746 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-ovndb-tls-certs\") pod \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.710967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsngq\" (UniqueName: \"kubernetes.io/projected/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-kube-api-access-tsngq\") pod \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.711077 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-httpd-config\") pod \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.711196 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-combined-ca-bundle\") pod \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.711372 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-config\") pod \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\" (UID: \"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb\") " Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.715521 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" (UID: "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.719268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-kube-api-access-tsngq" (OuterVolumeSpecName: "kube-api-access-tsngq") pod "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" (UID: "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb"). InnerVolumeSpecName "kube-api-access-tsngq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.790418 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" (UID: "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.791811 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-config" (OuterVolumeSpecName: "config") pod "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" (UID: "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.813522 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.813592 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsngq\" (UniqueName: \"kubernetes.io/projected/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-kube-api-access-tsngq\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.813608 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.813620 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.815212 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" (UID: "a168d2d9-a037-42f3-a4a3-6ad2bdec15eb"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:30 crc kubenswrapper[4758]: I1004 11:08:30.915166 4758 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.501550 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cf6c6588-6rzc8" event={"ID":"a168d2d9-a037-42f3-a4a3-6ad2bdec15eb","Type":"ContainerDied","Data":"f44166449473b09958e24799008ef9e8bde5ae7007bb57a303f04d64a7025c01"} Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.501854 4758 scope.go:117] "RemoveContainer" containerID="b9ce32a8fce8271bdc3fbe67f0eee4af8f3430da4115e44584099b07cd3170c3" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.501647 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cf6c6588-6rzc8" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.508387 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"49fbaf4c-21ec-4246-86b8-1361ae5f3149","Type":"ContainerStarted","Data":"ac3f9204f77b14d129e3b8cc7736472233de7226d7f8162f9491ea8b57fb765c"} Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.508690 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.513055 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerStarted","Data":"f337bb39476cb995331072b16661f4142fdf08429bea1a64f03d304f5b393c65"} Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.513218 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.528684 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66cf6c6588-6rzc8"] Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.534962 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66cf6c6588-6rzc8"] Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.541400 4758 scope.go:117] "RemoveContainer" containerID="7a57b136c7ec5a8825880144eddb8123052bc39b0a8532fc339e04bf6f56bd96" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.549925 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.027678358 podStartE2EDuration="5.54990721s" podCreationTimestamp="2025-10-04 11:08:26 +0000 UTC" firstStartedPulling="2025-10-04 11:08:27.37898785 +0000 UTC m=+1084.671638729" lastFinishedPulling="2025-10-04 11:08:30.901216702 +0000 UTC m=+1088.193867581" observedRunningTime="2025-10-04 11:08:31.542340085 +0000 UTC m=+1088.834990984" watchObservedRunningTime="2025-10-04 11:08:31.54990721 +0000 UTC m=+1088.842558109" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.574550 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.574527808 podStartE2EDuration="3.574527808s" podCreationTimestamp="2025-10-04 11:08:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:31.566338466 +0000 UTC m=+1088.858989385" watchObservedRunningTime="2025-10-04 11:08:31.574527808 +0000 UTC m=+1088.867178707" Oct 04 11:08:31 crc kubenswrapper[4758]: I1004 11:08:31.835785 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.207215 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.266914 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-ftr94"] Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.267162 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerName="dnsmasq-dns" containerID="cri-o://288db352cd496aa2fe0003f561f25e8ad21a7116b4f0d98b495aa33bacfa3c5d" gracePeriod=10 Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.269684 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.356389 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.534277 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerID="288db352cd496aa2fe0003f561f25e8ad21a7116b4f0d98b495aa33bacfa3c5d" exitCode=0 Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.534346 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" event={"ID":"cc84e061-4084-445a-b6ef-1e015ebd5007","Type":"ContainerDied","Data":"288db352cd496aa2fe0003f561f25e8ad21a7116b4f0d98b495aa33bacfa3c5d"} Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.541957 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="cinder-scheduler" containerID="cri-o://a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f" gracePeriod=30 Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.542127 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="probe" containerID="cri-o://c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000" gracePeriod=30 Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.881860 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.956970 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-swift-storage-0\") pod \"cc84e061-4084-445a-b6ef-1e015ebd5007\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.957034 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-config\") pod \"cc84e061-4084-445a-b6ef-1e015ebd5007\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.957163 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-svc\") pod \"cc84e061-4084-445a-b6ef-1e015ebd5007\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.957232 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmpk4\" (UniqueName: \"kubernetes.io/projected/cc84e061-4084-445a-b6ef-1e015ebd5007-kube-api-access-lmpk4\") pod \"cc84e061-4084-445a-b6ef-1e015ebd5007\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.957261 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-nb\") pod \"cc84e061-4084-445a-b6ef-1e015ebd5007\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " Oct 04 11:08:32 crc kubenswrapper[4758]: I1004 11:08:32.957292 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-sb\") pod \"cc84e061-4084-445a-b6ef-1e015ebd5007\" (UID: \"cc84e061-4084-445a-b6ef-1e015ebd5007\") " Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.006292 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc84e061-4084-445a-b6ef-1e015ebd5007-kube-api-access-lmpk4" (OuterVolumeSpecName: "kube-api-access-lmpk4") pod "cc84e061-4084-445a-b6ef-1e015ebd5007" (UID: "cc84e061-4084-445a-b6ef-1e015ebd5007"). InnerVolumeSpecName "kube-api-access-lmpk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.059195 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmpk4\" (UniqueName: \"kubernetes.io/projected/cc84e061-4084-445a-b6ef-1e015ebd5007-kube-api-access-lmpk4\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.171615 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cc84e061-4084-445a-b6ef-1e015ebd5007" (UID: "cc84e061-4084-445a-b6ef-1e015ebd5007"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.191156 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc84e061-4084-445a-b6ef-1e015ebd5007" (UID: "cc84e061-4084-445a-b6ef-1e015ebd5007"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.196499 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc84e061-4084-445a-b6ef-1e015ebd5007" (UID: "cc84e061-4084-445a-b6ef-1e015ebd5007"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.204730 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc84e061-4084-445a-b6ef-1e015ebd5007" (UID: "cc84e061-4084-445a-b6ef-1e015ebd5007"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.218459 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-config" (OuterVolumeSpecName: "config") pod "cc84e061-4084-445a-b6ef-1e015ebd5007" (UID: "cc84e061-4084-445a-b6ef-1e015ebd5007"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.262372 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.262399 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.262408 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.262416 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.262424 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc84e061-4084-445a-b6ef-1e015ebd5007-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.351464 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" path="/var/lib/kubelet/pods/a168d2d9-a037-42f3-a4a3-6ad2bdec15eb/volumes" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.551653 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" event={"ID":"cc84e061-4084-445a-b6ef-1e015ebd5007","Type":"ContainerDied","Data":"bad93bc421f7d3c5f49bcfe83220b933bd47e2162e67ceb8bf4c296bc9c9d0cf"} Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.551707 4758 scope.go:117] "RemoveContainer" containerID="288db352cd496aa2fe0003f561f25e8ad21a7116b4f0d98b495aa33bacfa3c5d" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.551725 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-ftr94" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.575932 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-ftr94"] Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.579033 4758 scope.go:117] "RemoveContainer" containerID="45f4c106cee77e1069cacb715c40e9166ffe8cd2de6978027982ec139f81a11b" Oct 04 11:08:33 crc kubenswrapper[4758]: I1004 11:08:33.583264 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-ftr94"] Oct 04 11:08:34 crc kubenswrapper[4758]: I1004 11:08:34.164477 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:34 crc kubenswrapper[4758]: I1004 11:08:34.192613 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8488664d88-ghdnd" Oct 04 11:08:34 crc kubenswrapper[4758]: I1004 11:08:34.570706 4758 generic.go:334] "Generic (PLEG): container finished" podID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerID="c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000" exitCode=0 Oct 04 11:08:34 crc kubenswrapper[4758]: I1004 11:08:34.570805 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6","Type":"ContainerDied","Data":"c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000"} Oct 04 11:08:34 crc kubenswrapper[4758]: I1004 11:08:34.765200 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.012926 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-ccf54cd5-74s2f" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.222959 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.322939 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data-custom\") pod \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323086 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data\") pod \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323158 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-etc-machine-id\") pod \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323211 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-combined-ca-bundle\") pod \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323240 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-scripts\") pod \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323283 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkffm\" (UniqueName: \"kubernetes.io/projected/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-kube-api-access-xkffm\") pod \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\" (UID: \"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6\") " Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323290 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" (UID: "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.323690 4758 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.330229 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-kube-api-access-xkffm" (OuterVolumeSpecName: "kube-api-access-xkffm") pod "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" (UID: "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6"). InnerVolumeSpecName "kube-api-access-xkffm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.345561 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-scripts" (OuterVolumeSpecName: "scripts") pod "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" (UID: "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.349806 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" (UID: "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.350445 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" path="/var/lib/kubelet/pods/cc84e061-4084-445a-b6ef-1e015ebd5007/volumes" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.377181 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" (UID: "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.425204 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.425655 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.425801 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkffm\" (UniqueName: \"kubernetes.io/projected/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-kube-api-access-xkffm\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.425815 4758 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.426230 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data" (OuterVolumeSpecName: "config-data") pod "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" (UID: "abfd6e4e-c59d-4029-aaa7-bee4e4c289d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.527009 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.581772 4758 generic.go:334] "Generic (PLEG): container finished" podID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerID="a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f" exitCode=0 Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.581839 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.581857 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6","Type":"ContainerDied","Data":"a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f"} Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.582250 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abfd6e4e-c59d-4029-aaa7-bee4e4c289d6","Type":"ContainerDied","Data":"0afa976774a895c7fe6880049cee38fda4b23bbb494259cabd02405b220d4ec5"} Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.582270 4758 scope.go:117] "RemoveContainer" containerID="c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.601566 4758 scope.go:117] "RemoveContainer" containerID="a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.617536 4758 scope.go:117] "RemoveContainer" containerID="c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.617927 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000\": container with ID starting with c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000 not found: ID does not exist" containerID="c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.617970 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000"} err="failed to get container status \"c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000\": rpc error: code = NotFound desc = could not find container \"c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000\": container with ID starting with c2e935245d82616f90c224d1bd28bf79d3dfd6c7274cc7534866b5be743a2000 not found: ID does not exist" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.617998 4758 scope.go:117] "RemoveContainer" containerID="a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.618375 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f\": container with ID starting with a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f not found: ID does not exist" containerID="a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.618477 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f"} err="failed to get container status \"a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f\": rpc error: code = NotFound desc = could not find container \"a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f\": container with ID starting with a7dc0b361642ef967fe4ee22030157e005ed079627c093ce241ac9c309c2211f not found: ID does not exist" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.623787 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.633420 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.648803 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.649140 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="probe" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649156 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="probe" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.649170 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-httpd" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649176 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-httpd" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.649190 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerName="init" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649196 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerName="init" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.649203 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerName="dnsmasq-dns" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649209 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerName="dnsmasq-dns" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.649225 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="cinder-scheduler" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649230 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="cinder-scheduler" Oct 04 11:08:35 crc kubenswrapper[4758]: E1004 11:08:35.649252 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-api" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649257 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-api" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649422 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="probe" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649435 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-httpd" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649454 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" containerName="cinder-scheduler" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649461 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a168d2d9-a037-42f3-a4a3-6ad2bdec15eb" containerName="neutron-api" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.649471 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc84e061-4084-445a-b6ef-1e015ebd5007" containerName="dnsmasq-dns" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.650389 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.653741 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.666012 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.729445 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.729524 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.729554 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.729568 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64chm\" (UniqueName: \"kubernetes.io/projected/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-kube-api-access-64chm\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.729598 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-config-data\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.729633 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-scripts\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.831543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.831609 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.831637 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.831654 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64chm\" (UniqueName: \"kubernetes.io/projected/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-kube-api-access-64chm\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.831683 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-config-data\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.831718 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-scripts\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.832365 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.836319 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-config-data\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.837184 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.837231 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-scripts\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.842378 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.852583 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64chm\" (UniqueName: \"kubernetes.io/projected/05d625cd-c9ec-4b81-a929-e1b9027e8c8e-kube-api-access-64chm\") pod \"cinder-scheduler-0\" (UID: \"05d625cd-c9ec-4b81-a929-e1b9027e8c8e\") " pod="openstack/cinder-scheduler-0" Oct 04 11:08:35 crc kubenswrapper[4758]: I1004 11:08:35.965120 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 11:08:36 crc kubenswrapper[4758]: I1004 11:08:36.441979 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 11:08:36 crc kubenswrapper[4758]: I1004 11:08:36.602122 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05d625cd-c9ec-4b81-a929-e1b9027e8c8e","Type":"ContainerStarted","Data":"1bfa96d1735ce797d0c4b2047db5964368327baa56ffc0be20993c28f3f3c6ac"} Oct 04 11:08:37 crc kubenswrapper[4758]: I1004 11:08:37.342955 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abfd6e4e-c59d-4029-aaa7-bee4e4c289d6" path="/var/lib/kubelet/pods/abfd6e4e-c59d-4029-aaa7-bee4e4c289d6/volumes" Oct 04 11:08:37 crc kubenswrapper[4758]: I1004 11:08:37.611701 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05d625cd-c9ec-4b81-a929-e1b9027e8c8e","Type":"ContainerStarted","Data":"3b78550d61944bed825babf33d7c9888e9f9f255505c309d6f3a8eab8edb3dd6"} Oct 04 11:08:38 crc kubenswrapper[4758]: I1004 11:08:38.619782 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05d625cd-c9ec-4b81-a929-e1b9027e8c8e","Type":"ContainerStarted","Data":"f39cc4d25f44f929022727b1d5c378f910b0a67e1f3413d7eea35877171b43e3"} Oct 04 11:08:38 crc kubenswrapper[4758]: I1004 11:08:38.643938 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.6439229170000003 podStartE2EDuration="3.643922917s" podCreationTimestamp="2025-10-04 11:08:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:38.635765716 +0000 UTC m=+1095.928416605" watchObservedRunningTime="2025-10-04 11:08:38.643922917 +0000 UTC m=+1095.936573806" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.011290 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.012638 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.015582 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.015630 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-j7bxw" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.016447 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.028886 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.096199 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78723cfd-8fe5-4c7c-9add-90ac24e14f55-combined-ca-bundle\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.096268 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvbjz\" (UniqueName: \"kubernetes.io/projected/78723cfd-8fe5-4c7c-9add-90ac24e14f55-kube-api-access-kvbjz\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.096407 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78723cfd-8fe5-4c7c-9add-90ac24e14f55-openstack-config-secret\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.096444 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78723cfd-8fe5-4c7c-9add-90ac24e14f55-openstack-config\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.197592 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78723cfd-8fe5-4c7c-9add-90ac24e14f55-openstack-config-secret\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.197629 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78723cfd-8fe5-4c7c-9add-90ac24e14f55-openstack-config\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.198512 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/78723cfd-8fe5-4c7c-9add-90ac24e14f55-openstack-config\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.197666 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78723cfd-8fe5-4c7c-9add-90ac24e14f55-combined-ca-bundle\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.198592 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvbjz\" (UniqueName: \"kubernetes.io/projected/78723cfd-8fe5-4c7c-9add-90ac24e14f55-kube-api-access-kvbjz\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.204247 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/78723cfd-8fe5-4c7c-9add-90ac24e14f55-openstack-config-secret\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.216314 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78723cfd-8fe5-4c7c-9add-90ac24e14f55-combined-ca-bundle\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.218614 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvbjz\" (UniqueName: \"kubernetes.io/projected/78723cfd-8fe5-4c7c-9add-90ac24e14f55-kube-api-access-kvbjz\") pod \"openstackclient\" (UID: \"78723cfd-8fe5-4c7c-9add-90ac24e14f55\") " pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.341061 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 11:08:39 crc kubenswrapper[4758]: I1004 11:08:39.857182 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 11:08:40 crc kubenswrapper[4758]: I1004 11:08:40.638924 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"78723cfd-8fe5-4c7c-9add-90ac24e14f55","Type":"ContainerStarted","Data":"d370966c7447222a77d85186430752df3c447a649c68fab4ec0f1ca9325348c4"} Oct 04 11:08:40 crc kubenswrapper[4758]: I1004 11:08:40.965750 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 11:08:41 crc kubenswrapper[4758]: I1004 11:08:41.836411 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.247994 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.248957 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-central-agent" containerID="cri-o://bd4f234257fdd115195d774554c71ecbc04fc9d271dbcd85d8680acbf29be583" gracePeriod=30 Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.249128 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="proxy-httpd" containerID="cri-o://f337bb39476cb995331072b16661f4142fdf08429bea1a64f03d304f5b393c65" gracePeriod=30 Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.249224 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="sg-core" containerID="cri-o://f9186911306d1a73e24f8f65029c47e948fc6f1de39edc8dcb9c0e9e53f5f473" gracePeriod=30 Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.249273 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-notification-agent" containerID="cri-o://5295b93ea52879cc60ec75d6f7a945518e30b60d3f4f500516f5cb8ddc5f7e91" gracePeriod=30 Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.261773 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.163:3000/\": EOF" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.435803 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-996d997c5-pbjdp"] Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.438240 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.440758 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.440821 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.442931 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.459630 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-996d997c5-pbjdp"] Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509661 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-combined-ca-bundle\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509747 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-etc-swift\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509807 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-internal-tls-certs\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509869 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj5g8\" (UniqueName: \"kubernetes.io/projected/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-kube-api-access-zj5g8\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509909 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-run-httpd\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509943 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-public-tls-certs\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.509985 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-log-httpd\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.510017 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-config-data\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.611975 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-run-httpd\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612031 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-public-tls-certs\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612067 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-log-httpd\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612093 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-config-data\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-combined-ca-bundle\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612183 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-etc-swift\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612220 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-internal-tls-certs\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612258 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj5g8\" (UniqueName: \"kubernetes.io/projected/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-kube-api-access-zj5g8\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.612558 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-run-httpd\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.613400 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-log-httpd\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.619611 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-config-data\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.622058 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-etc-swift\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.623768 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-internal-tls-certs\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.628397 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-public-tls-certs\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.631941 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj5g8\" (UniqueName: \"kubernetes.io/projected/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-kube-api-access-zj5g8\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.639103 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9170c3d7-3b7f-4945-9e95-76c4b9d476f0-combined-ca-bundle\") pod \"swift-proxy-996d997c5-pbjdp\" (UID: \"9170c3d7-3b7f-4945-9e95-76c4b9d476f0\") " pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.684792 4758 generic.go:334] "Generic (PLEG): container finished" podID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerID="f337bb39476cb995331072b16661f4142fdf08429bea1a64f03d304f5b393c65" exitCode=0 Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.684825 4758 generic.go:334] "Generic (PLEG): container finished" podID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerID="f9186911306d1a73e24f8f65029c47e948fc6f1de39edc8dcb9c0e9e53f5f473" exitCode=2 Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.684850 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerDied","Data":"f337bb39476cb995331072b16661f4142fdf08429bea1a64f03d304f5b393c65"} Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.684880 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerDied","Data":"f9186911306d1a73e24f8f65029c47e948fc6f1de39edc8dcb9c0e9e53f5f473"} Oct 04 11:08:44 crc kubenswrapper[4758]: I1004 11:08:44.753760 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:45 crc kubenswrapper[4758]: I1004 11:08:45.336794 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-996d997c5-pbjdp"] Oct 04 11:08:45 crc kubenswrapper[4758]: I1004 11:08:45.706145 4758 generic.go:334] "Generic (PLEG): container finished" podID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerID="bd4f234257fdd115195d774554c71ecbc04fc9d271dbcd85d8680acbf29be583" exitCode=0 Oct 04 11:08:45 crc kubenswrapper[4758]: I1004 11:08:45.706548 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerDied","Data":"bd4f234257fdd115195d774554c71ecbc04fc9d271dbcd85d8680acbf29be583"} Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.270531 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.305698 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hw2fc"] Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.306932 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.322956 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hw2fc"] Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.417726 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-8tnvw"] Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.419675 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.468998 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgbfb\" (UniqueName: \"kubernetes.io/projected/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579-kube-api-access-vgbfb\") pod \"nova-api-db-create-hw2fc\" (UID: \"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579\") " pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.469477 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8tnvw"] Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.570251 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgbfb\" (UniqueName: \"kubernetes.io/projected/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579-kube-api-access-vgbfb\") pod \"nova-api-db-create-hw2fc\" (UID: \"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579\") " pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.570417 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpkdr\" (UniqueName: \"kubernetes.io/projected/53b6351b-2066-44a5-90fe-f06e671b9452-kube-api-access-tpkdr\") pod \"nova-cell0-db-create-8tnvw\" (UID: \"53b6351b-2066-44a5-90fe-f06e671b9452\") " pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.592700 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgbfb\" (UniqueName: \"kubernetes.io/projected/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579-kube-api-access-vgbfb\") pod \"nova-api-db-create-hw2fc\" (UID: \"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579\") " pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.618732 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-s9zzx"] Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.619846 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.633344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-s9zzx"] Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.636576 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.671832 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpkdr\" (UniqueName: \"kubernetes.io/projected/53b6351b-2066-44a5-90fe-f06e671b9452-kube-api-access-tpkdr\") pod \"nova-cell0-db-create-8tnvw\" (UID: \"53b6351b-2066-44a5-90fe-f06e671b9452\") " pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.706803 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpkdr\" (UniqueName: \"kubernetes.io/projected/53b6351b-2066-44a5-90fe-f06e671b9452-kube-api-access-tpkdr\") pod \"nova-cell0-db-create-8tnvw\" (UID: \"53b6351b-2066-44a5-90fe-f06e671b9452\") " pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.715254 4758 generic.go:334] "Generic (PLEG): container finished" podID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerID="5295b93ea52879cc60ec75d6f7a945518e30b60d3f4f500516f5cb8ddc5f7e91" exitCode=0 Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.715288 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerDied","Data":"5295b93ea52879cc60ec75d6f7a945518e30b60d3f4f500516f5cb8ddc5f7e91"} Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.773923 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5p6v\" (UniqueName: \"kubernetes.io/projected/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6-kube-api-access-b5p6v\") pod \"nova-cell1-db-create-s9zzx\" (UID: \"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6\") " pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.783943 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.878907 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5p6v\" (UniqueName: \"kubernetes.io/projected/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6-kube-api-access-b5p6v\") pod \"nova-cell1-db-create-s9zzx\" (UID: \"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6\") " pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.901685 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5p6v\" (UniqueName: \"kubernetes.io/projected/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6-kube-api-access-b5p6v\") pod \"nova-cell1-db-create-s9zzx\" (UID: \"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6\") " pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:46 crc kubenswrapper[4758]: I1004 11:08:46.959252 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:49 crc kubenswrapper[4758]: I1004 11:08:49.745581 4758 generic.go:334] "Generic (PLEG): container finished" podID="61f721a8-00fb-446c-8e98-922827d2fc39" containerID="1879f1801937d4659da7d57131cbee347b49df5d138e811f68d0a8385f43bfe5" exitCode=137 Oct 04 11:08:49 crc kubenswrapper[4758]: I1004 11:08:49.745677 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c98cb656b-lczgq" event={"ID":"61f721a8-00fb-446c-8e98-922827d2fc39","Type":"ContainerDied","Data":"1879f1801937d4659da7d57131cbee347b49df5d138e811f68d0a8385f43bfe5"} Oct 04 11:08:51 crc kubenswrapper[4758]: W1004 11:08:51.733276 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9170c3d7_3b7f_4945_9e95_76c4b9d476f0.slice/crio-9fd4bd3b0dbf569af07ecb5021a3490ce768cee53739428ebe5b2daf06e28995 WatchSource:0}: Error finding container 9fd4bd3b0dbf569af07ecb5021a3490ce768cee53739428ebe5b2daf06e28995: Status 404 returned error can't find the container with id 9fd4bd3b0dbf569af07ecb5021a3490ce768cee53739428ebe5b2daf06e28995 Oct 04 11:08:51 crc kubenswrapper[4758]: I1004 11:08:51.807502 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-996d997c5-pbjdp" event={"ID":"9170c3d7-3b7f-4945-9e95-76c4b9d476f0","Type":"ContainerStarted","Data":"9fd4bd3b0dbf569af07ecb5021a3490ce768cee53739428ebe5b2daf06e28995"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.229126 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.411682 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-sg-core-conf-yaml\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.411756 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-log-httpd\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.411792 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-config-data\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.411856 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj7r6\" (UniqueName: \"kubernetes.io/projected/517aff26-ac5d-4395-9d0c-0d14690b7684-kube-api-access-nj7r6\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.411942 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-scripts\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.411993 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-combined-ca-bundle\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.412015 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-run-httpd\") pod \"517aff26-ac5d-4395-9d0c-0d14690b7684\" (UID: \"517aff26-ac5d-4395-9d0c-0d14690b7684\") " Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.412771 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.412972 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.422301 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-scripts" (OuterVolumeSpecName: "scripts") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.422449 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/517aff26-ac5d-4395-9d0c-0d14690b7684-kube-api-access-nj7r6" (OuterVolumeSpecName: "kube-api-access-nj7r6") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "kube-api-access-nj7r6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.426189 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8tnvw"] Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.433422 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-s9zzx"] Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.454263 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.518678 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.518977 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.518988 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/517aff26-ac5d-4395-9d0c-0d14690b7684-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.518997 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj7r6\" (UniqueName: \"kubernetes.io/projected/517aff26-ac5d-4395-9d0c-0d14690b7684-kube-api-access-nj7r6\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.519006 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.539322 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.550524 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hw2fc"] Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.621370 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.621428 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-config-data" (OuterVolumeSpecName: "config-data") pod "517aff26-ac5d-4395-9d0c-0d14690b7684" (UID: "517aff26-ac5d-4395-9d0c-0d14690b7684"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.722884 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517aff26-ac5d-4395-9d0c-0d14690b7684-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.817712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"78723cfd-8fe5-4c7c-9add-90ac24e14f55","Type":"ContainerStarted","Data":"e75bba235cd0b12876ee2593730878820e54517b01b4099b0ef1f6f74375e959"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.819829 4758 generic.go:334] "Generic (PLEG): container finished" podID="53b6351b-2066-44a5-90fe-f06e671b9452" containerID="ad746931127bfee5948e7431d4b3fbb488edcc36ffbecaf167f89614124c4f3b" exitCode=0 Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.819877 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8tnvw" event={"ID":"53b6351b-2066-44a5-90fe-f06e671b9452","Type":"ContainerDied","Data":"ad746931127bfee5948e7431d4b3fbb488edcc36ffbecaf167f89614124c4f3b"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.819898 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8tnvw" event={"ID":"53b6351b-2066-44a5-90fe-f06e671b9452","Type":"ContainerStarted","Data":"d9d843d0467bfb3a0861f022e83ad06b565a9104b614d5c00656abbc7a0d8c30"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.826672 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-996d997c5-pbjdp" event={"ID":"9170c3d7-3b7f-4945-9e95-76c4b9d476f0","Type":"ContainerStarted","Data":"ce963de6157ebcf045c00da17e498bb758518c10d53d93dc29482e454ac42c73"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.826708 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-996d997c5-pbjdp" event={"ID":"9170c3d7-3b7f-4945-9e95-76c4b9d476f0","Type":"ContainerStarted","Data":"65a0b5fad85cd438452fc5253a754860b547aa49195581b8c7912d98e343f8dc"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.827055 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.827111 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.828757 4758 generic.go:334] "Generic (PLEG): container finished" podID="b6f9b2b3-f0ed-4fb8-837f-421ebaf92579" containerID="13263f9591ec9b17b5cce4b8a589f37bb71f6bd905aec33d3881ef7c493d85b6" exitCode=0 Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.828804 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hw2fc" event={"ID":"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579","Type":"ContainerDied","Data":"13263f9591ec9b17b5cce4b8a589f37bb71f6bd905aec33d3881ef7c493d85b6"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.828822 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hw2fc" event={"ID":"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579","Type":"ContainerStarted","Data":"60ff676fed06d5cf9f48daded9b300a4fa28b0eb7237282eb6df2e614e9a7a94"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.832370 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c98cb656b-lczgq" event={"ID":"61f721a8-00fb-446c-8e98-922827d2fc39","Type":"ContainerStarted","Data":"907c20f8d789d88d77b050f300f43fc0524313fe866788751b6dc353ea432f44"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.834688 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.593978314 podStartE2EDuration="14.834661797s" podCreationTimestamp="2025-10-04 11:08:38 +0000 UTC" firstStartedPulling="2025-10-04 11:08:39.863747079 +0000 UTC m=+1097.156397968" lastFinishedPulling="2025-10-04 11:08:52.104430562 +0000 UTC m=+1109.397081451" observedRunningTime="2025-10-04 11:08:52.830610297 +0000 UTC m=+1110.123261196" watchObservedRunningTime="2025-10-04 11:08:52.834661797 +0000 UTC m=+1110.127312686" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.835904 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"517aff26-ac5d-4395-9d0c-0d14690b7684","Type":"ContainerDied","Data":"e714edbfd8b803163a5d1d981dcf115bdeccdefd2956befb8406e0b16ba8bee0"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.835934 4758 scope.go:117] "RemoveContainer" containerID="f337bb39476cb995331072b16661f4142fdf08429bea1a64f03d304f5b393c65" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.836026 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.849824 4758 generic.go:334] "Generic (PLEG): container finished" podID="6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6" containerID="75a4965a9e350f492a8524fe9c228a230ff9b1601d74768f07ac8063c6626ac6" exitCode=0 Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.849869 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9zzx" event={"ID":"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6","Type":"ContainerDied","Data":"75a4965a9e350f492a8524fe9c228a230ff9b1601d74768f07ac8063c6626ac6"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.849900 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9zzx" event={"ID":"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6","Type":"ContainerStarted","Data":"7f9394229732c96120a63c0edeb30605e85e54318748a82d33134dfc3c8c842f"} Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.896014 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-996d997c5-pbjdp" podStartSLOduration=8.895997353 podStartE2EDuration="8.895997353s" podCreationTimestamp="2025-10-04 11:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:08:52.879420725 +0000 UTC m=+1110.172071614" watchObservedRunningTime="2025-10-04 11:08:52.895997353 +0000 UTC m=+1110.188648232" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.976723 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.978490 4758 scope.go:117] "RemoveContainer" containerID="f9186911306d1a73e24f8f65029c47e948fc6f1de39edc8dcb9c0e9e53f5f473" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.988504 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.995648 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:52 crc kubenswrapper[4758]: E1004 11:08:52.996051 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="proxy-httpd" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.996063 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="proxy-httpd" Oct 04 11:08:52 crc kubenswrapper[4758]: E1004 11:08:52.996075 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-notification-agent" Oct 04 11:08:52 crc kubenswrapper[4758]: I1004 11:08:52.996081 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-notification-agent" Oct 04 11:08:53 crc kubenswrapper[4758]: E1004 11:08:53.007682 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="sg-core" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.007708 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="sg-core" Oct 04 11:08:53 crc kubenswrapper[4758]: E1004 11:08:53.007744 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-central-agent" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.007750 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-central-agent" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.008076 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-central-agent" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.008095 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="sg-core" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.008125 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="ceilometer-notification-agent" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.008138 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" containerName="proxy-httpd" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.009619 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.014476 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.015778 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.016071 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.051833 4758 scope.go:117] "RemoveContainer" containerID="5295b93ea52879cc60ec75d6f7a945518e30b60d3f4f500516f5cb8ddc5f7e91" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130272 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130307 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx8xd\" (UniqueName: \"kubernetes.io/projected/92d53f6f-4eee-4ea4-8e0a-20daac629083-kube-api-access-nx8xd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130352 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-log-httpd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130372 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-run-httpd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130425 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130439 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-scripts\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.130482 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-config-data\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.135967 4758 scope.go:117] "RemoveContainer" containerID="bd4f234257fdd115195d774554c71ecbc04fc9d271dbcd85d8680acbf29be583" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.232421 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-config-data\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.233489 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.233595 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx8xd\" (UniqueName: \"kubernetes.io/projected/92d53f6f-4eee-4ea4-8e0a-20daac629083-kube-api-access-nx8xd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.233719 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-log-httpd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.233855 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-run-httpd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.234004 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.234137 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-scripts\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.234213 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-log-httpd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.234300 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-run-httpd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.238904 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-config-data\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.239870 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-scripts\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.240605 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.242668 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.256287 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx8xd\" (UniqueName: \"kubernetes.io/projected/92d53f6f-4eee-4ea4-8e0a-20daac629083-kube-api-access-nx8xd\") pod \"ceilometer-0\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.338153 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="517aff26-ac5d-4395-9d0c-0d14690b7684" path="/var/lib/kubelet/pods/517aff26-ac5d-4395-9d0c-0d14690b7684/volumes" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.340702 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:08:53 crc kubenswrapper[4758]: I1004 11:08:53.872046 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.198638 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.203161 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-log" containerID="cri-o://9c1625a48956f88788991172d2be04f15d991d81c49ef25582359fdbff2fb32d" gracePeriod=30 Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.203298 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-httpd" containerID="cri-o://c505624a2c304042dbfb72ea933c5983844a1a26b39789dcdaf77e0eb5c59df2" gracePeriod=30 Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.408496 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.460290 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpkdr\" (UniqueName: \"kubernetes.io/projected/53b6351b-2066-44a5-90fe-f06e671b9452-kube-api-access-tpkdr\") pod \"53b6351b-2066-44a5-90fe-f06e671b9452\" (UID: \"53b6351b-2066-44a5-90fe-f06e671b9452\") " Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.469242 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b6351b-2066-44a5-90fe-f06e671b9452-kube-api-access-tpkdr" (OuterVolumeSpecName: "kube-api-access-tpkdr") pod "53b6351b-2066-44a5-90fe-f06e671b9452" (UID: "53b6351b-2066-44a5-90fe-f06e671b9452"). InnerVolumeSpecName "kube-api-access-tpkdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.501465 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.516674 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.563501 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5p6v\" (UniqueName: \"kubernetes.io/projected/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6-kube-api-access-b5p6v\") pod \"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6\" (UID: \"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6\") " Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.563592 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgbfb\" (UniqueName: \"kubernetes.io/projected/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579-kube-api-access-vgbfb\") pod \"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579\" (UID: \"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579\") " Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.564066 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpkdr\" (UniqueName: \"kubernetes.io/projected/53b6351b-2066-44a5-90fe-f06e671b9452-kube-api-access-tpkdr\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.570652 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579-kube-api-access-vgbfb" (OuterVolumeSpecName: "kube-api-access-vgbfb") pod "b6f9b2b3-f0ed-4fb8-837f-421ebaf92579" (UID: "b6f9b2b3-f0ed-4fb8-837f-421ebaf92579"). InnerVolumeSpecName "kube-api-access-vgbfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.572238 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6-kube-api-access-b5p6v" (OuterVolumeSpecName: "kube-api-access-b5p6v") pod "6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6" (UID: "6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6"). InnerVolumeSpecName "kube-api-access-b5p6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.669330 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5p6v\" (UniqueName: \"kubernetes.io/projected/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6-kube-api-access-b5p6v\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.669369 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgbfb\" (UniqueName: \"kubernetes.io/projected/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579-kube-api-access-vgbfb\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.867688 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8tnvw" event={"ID":"53b6351b-2066-44a5-90fe-f06e671b9452","Type":"ContainerDied","Data":"d9d843d0467bfb3a0861f022e83ad06b565a9104b614d5c00656abbc7a0d8c30"} Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.867790 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9d843d0467bfb3a0861f022e83ad06b565a9104b614d5c00656abbc7a0d8c30" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.867913 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8tnvw" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.882396 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerStarted","Data":"6f8f9aa39bf4e0f9918645575c022dce250e9ca8cf631068777b1e92c68b1fce"} Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.883678 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerStarted","Data":"fd49949d2ebc0e613584f0607fae9fde577c3043d5c7836ad208c137112afffb"} Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.894414 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hw2fc" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.894517 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hw2fc" event={"ID":"b6f9b2b3-f0ed-4fb8-837f-421ebaf92579","Type":"ContainerDied","Data":"60ff676fed06d5cf9f48daded9b300a4fa28b0eb7237282eb6df2e614e9a7a94"} Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.895406 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60ff676fed06d5cf9f48daded9b300a4fa28b0eb7237282eb6df2e614e9a7a94" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.906482 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9zzx" event={"ID":"6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6","Type":"ContainerDied","Data":"7f9394229732c96120a63c0edeb30605e85e54318748a82d33134dfc3c8c842f"} Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.906597 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f9394229732c96120a63c0edeb30605e85e54318748a82d33134dfc3c8c842f" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.906709 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9zzx" Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.933350 4758 generic.go:334] "Generic (PLEG): container finished" podID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerID="9c1625a48956f88788991172d2be04f15d991d81c49ef25582359fdbff2fb32d" exitCode=143 Oct 04 11:08:54 crc kubenswrapper[4758]: I1004 11:08:54.933437 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d059c9e-fa03-46dd-b113-de161b74ecea","Type":"ContainerDied","Data":"9c1625a48956f88788991172d2be04f15d991d81c49ef25582359fdbff2fb32d"} Oct 04 11:08:55 crc kubenswrapper[4758]: I1004 11:08:55.973390 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerStarted","Data":"efa50282505e3dcde99e1070031fc0f6d76ac103d15def812dfbb6bcb67342c6"} Oct 04 11:08:57 crc kubenswrapper[4758]: I1004 11:08:57.015905 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerStarted","Data":"d79da546ff3f3946d480c3b033c65d953b200d453e8aae9a6eb7e5b1ab370dfd"} Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.024682 4758 generic.go:334] "Generic (PLEG): container finished" podID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerID="c505624a2c304042dbfb72ea933c5983844a1a26b39789dcdaf77e0eb5c59df2" exitCode=0 Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.024721 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d059c9e-fa03-46dd-b113-de161b74ecea","Type":"ContainerDied","Data":"c505624a2c304042dbfb72ea933c5983844a1a26b39789dcdaf77e0eb5c59df2"} Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.671839 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.747768 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.747828 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-logs\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.747849 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vr4m\" (UniqueName: \"kubernetes.io/projected/3d059c9e-fa03-46dd-b113-de161b74ecea-kube-api-access-9vr4m\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.747918 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-public-tls-certs\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.747963 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-httpd-run\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.748001 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-scripts\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.748034 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-config-data\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.748051 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-combined-ca-bundle\") pod \"3d059c9e-fa03-46dd-b113-de161b74ecea\" (UID: \"3d059c9e-fa03-46dd-b113-de161b74ecea\") " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.750227 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-logs" (OuterVolumeSpecName: "logs") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.753325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.755248 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.759288 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.761376 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-scripts" (OuterVolumeSpecName: "scripts") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.767617 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d059c9e-fa03-46dd-b113-de161b74ecea-kube-api-access-9vr4m" (OuterVolumeSpecName: "kube-api-access-9vr4m") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "kube-api-access-9vr4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.777619 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.825263 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.850646 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.850677 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.850685 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.850704 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.850714 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d059c9e-fa03-46dd-b113-de161b74ecea-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.850723 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vr4m\" (UniqueName: \"kubernetes.io/projected/3d059c9e-fa03-46dd-b113-de161b74ecea-kube-api-access-9vr4m\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.886316 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.887077 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-config-data" (OuterVolumeSpecName: "config-data") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.921077 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3d059c9e-fa03-46dd-b113-de161b74ecea" (UID: "3d059c9e-fa03-46dd-b113-de161b74ecea"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.953342 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.953577 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:58 crc kubenswrapper[4758]: I1004 11:08:58.953656 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d059c9e-fa03-46dd-b113-de161b74ecea-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.032661 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d059c9e-fa03-46dd-b113-de161b74ecea","Type":"ContainerDied","Data":"ba398e1bf14f52ec117196a7eb831cf40141bea2d9dddba248ca36eadf6ca4fc"} Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.032920 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.033647 4758 scope.go:117] "RemoveContainer" containerID="c505624a2c304042dbfb72ea933c5983844a1a26b39789dcdaf77e0eb5c59df2" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.037837 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerStarted","Data":"9209fc34d8a144b1d61f51ddfba3f65281665b1eb154215091660ba780d10b2e"} Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.037874 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.053064 4758 scope.go:117] "RemoveContainer" containerID="9c1625a48956f88788991172d2be04f15d991d81c49ef25582359fdbff2fb32d" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.061957 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.785528766 podStartE2EDuration="7.061937599s" podCreationTimestamp="2025-10-04 11:08:52 +0000 UTC" firstStartedPulling="2025-10-04 11:08:53.888137609 +0000 UTC m=+1111.180788498" lastFinishedPulling="2025-10-04 11:08:58.164546442 +0000 UTC m=+1115.457197331" observedRunningTime="2025-10-04 11:08:59.059273757 +0000 UTC m=+1116.351924646" watchObservedRunningTime="2025-10-04 11:08:59.061937599 +0000 UTC m=+1116.354588488" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.095720 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.110298 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120267 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:08:59 crc kubenswrapper[4758]: E1004 11:08:59.120671 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-httpd" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120692 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-httpd" Oct 04 11:08:59 crc kubenswrapper[4758]: E1004 11:08:59.120710 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120718 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: E1004 11:08:59.120732 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-log" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120739 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-log" Oct 04 11:08:59 crc kubenswrapper[4758]: E1004 11:08:59.120754 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f9b2b3-f0ed-4fb8-837f-421ebaf92579" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120762 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f9b2b3-f0ed-4fb8-837f-421ebaf92579" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: E1004 11:08:59.120784 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b6351b-2066-44a5-90fe-f06e671b9452" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120792 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b6351b-2066-44a5-90fe-f06e671b9452" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.120969 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f9b2b3-f0ed-4fb8-837f-421ebaf92579" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.121006 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.121020 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-httpd" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.121033 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" containerName="glance-log" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.121049 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b6351b-2066-44a5-90fe-f06e671b9452" containerName="mariadb-database-create" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.122142 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.125655 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.125948 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.135031 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261368 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7359e803-55a1-4586-bd84-5b389daf4167-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261411 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-scripts\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261456 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261481 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d27lz\" (UniqueName: \"kubernetes.io/projected/7359e803-55a1-4586-bd84-5b389daf4167-kube-api-access-d27lz\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261520 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-config-data\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261597 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261615 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7359e803-55a1-4586-bd84-5b389daf4167-logs\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.261658 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.337422 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d059c9e-fa03-46dd-b113-de161b74ecea" path="/var/lib/kubelet/pods/3d059c9e-fa03-46dd-b113-de161b74ecea/volumes" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.363437 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.363516 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7359e803-55a1-4586-bd84-5b389daf4167-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.363924 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-scripts\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.364154 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.364204 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d27lz\" (UniqueName: \"kubernetes.io/projected/7359e803-55a1-4586-bd84-5b389daf4167-kube-api-access-d27lz\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.364314 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-config-data\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.364378 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.364398 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7359e803-55a1-4586-bd84-5b389daf4167-logs\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.364891 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.365439 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7359e803-55a1-4586-bd84-5b389daf4167-logs\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.365786 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7359e803-55a1-4586-bd84-5b389daf4167-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.368960 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.371305 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-config-data\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.374253 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-scripts\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.374894 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7359e803-55a1-4586-bd84-5b389daf4167-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.382622 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d27lz\" (UniqueName: \"kubernetes.io/projected/7359e803-55a1-4586-bd84-5b389daf4167-kube-api-access-d27lz\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.398848 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"7359e803-55a1-4586-bd84-5b389daf4167\") " pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.436193 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.436416 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-log" containerID="cri-o://190b414813fbd25927e7ab3d30a73c8c1449cd4df4fb8e8945f3c6f5e3b9cbfb" gracePeriod=30 Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.436596 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-httpd" containerID="cri-o://4ab23623bbd35983978197a60e4cbe69801b418dee40c090e20920c56ad37e18" gracePeriod=30 Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.448269 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.795702 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:08:59 crc kubenswrapper[4758]: I1004 11:08:59.796674 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-996d997c5-pbjdp" Oct 04 11:09:00 crc kubenswrapper[4758]: I1004 11:09:00.045832 4758 generic.go:334] "Generic (PLEG): container finished" podID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerID="190b414813fbd25927e7ab3d30a73c8c1449cd4df4fb8e8945f3c6f5e3b9cbfb" exitCode=143 Oct 04 11:09:00 crc kubenswrapper[4758]: I1004 11:09:00.045911 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6ad8010-4207-4d04-9765-5b81facf9cad","Type":"ContainerDied","Data":"190b414813fbd25927e7ab3d30a73c8c1449cd4df4fb8e8945f3c6f5e3b9cbfb"} Oct 04 11:09:00 crc kubenswrapper[4758]: I1004 11:09:00.183348 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.057931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7359e803-55a1-4586-bd84-5b389daf4167","Type":"ContainerStarted","Data":"e868fd9e1f6010d1c738d06b385dfc46ee914619896dbdc1d446d7eaf2bcd553"} Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.058421 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7359e803-55a1-4586-bd84-5b389daf4167","Type":"ContainerStarted","Data":"f038252ec2f96d196b5111555ae33dcb901054594fa480e2b2a6b7934aa8f5e7"} Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.187418 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.187663 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-central-agent" containerID="cri-o://6f8f9aa39bf4e0f9918645575c022dce250e9ca8cf631068777b1e92c68b1fce" gracePeriod=30 Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.187779 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-notification-agent" containerID="cri-o://efa50282505e3dcde99e1070031fc0f6d76ac103d15def812dfbb6bcb67342c6" gracePeriod=30 Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.187817 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="sg-core" containerID="cri-o://d79da546ff3f3946d480c3b033c65d953b200d453e8aae9a6eb7e5b1ab370dfd" gracePeriod=30 Oct 04 11:09:01 crc kubenswrapper[4758]: I1004 11:09:01.187843 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="proxy-httpd" containerID="cri-o://9209fc34d8a144b1d61f51ddfba3f65281665b1eb154215091660ba780d10b2e" gracePeriod=30 Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.068922 4758 generic.go:334] "Generic (PLEG): container finished" podID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerID="9209fc34d8a144b1d61f51ddfba3f65281665b1eb154215091660ba780d10b2e" exitCode=0 Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.069246 4758 generic.go:334] "Generic (PLEG): container finished" podID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerID="d79da546ff3f3946d480c3b033c65d953b200d453e8aae9a6eb7e5b1ab370dfd" exitCode=2 Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.069255 4758 generic.go:334] "Generic (PLEG): container finished" podID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerID="efa50282505e3dcde99e1070031fc0f6d76ac103d15def812dfbb6bcb67342c6" exitCode=0 Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.069115 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerDied","Data":"9209fc34d8a144b1d61f51ddfba3f65281665b1eb154215091660ba780d10b2e"} Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.069318 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerDied","Data":"d79da546ff3f3946d480c3b033c65d953b200d453e8aae9a6eb7e5b1ab370dfd"} Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.069331 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerDied","Data":"efa50282505e3dcde99e1070031fc0f6d76ac103d15def812dfbb6bcb67342c6"} Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.071371 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7359e803-55a1-4586-bd84-5b389daf4167","Type":"ContainerStarted","Data":"826afe6165f477abd0bc6763dab025e99342ac489ea417bb546ea3589aa520f5"} Oct 04 11:09:02 crc kubenswrapper[4758]: I1004 11:09:02.094532 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.094517302 podStartE2EDuration="3.094517302s" podCreationTimestamp="2025-10-04 11:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:02.089005573 +0000 UTC m=+1119.381656462" watchObservedRunningTime="2025-10-04 11:09:02.094517302 +0000 UTC m=+1119.387168191" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.087670 4758 generic.go:334] "Generic (PLEG): container finished" podID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerID="4ab23623bbd35983978197a60e4cbe69801b418dee40c090e20920c56ad37e18" exitCode=0 Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.087738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6ad8010-4207-4d04-9765-5b81facf9cad","Type":"ContainerDied","Data":"4ab23623bbd35983978197a60e4cbe69801b418dee40c090e20920c56ad37e18"} Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.447198 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.534719 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.534812 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-combined-ca-bundle\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.535829 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-logs\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.535866 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-internal-tls-certs\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.535902 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-httpd-run\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.535970 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-config-data\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.536018 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmwhc\" (UniqueName: \"kubernetes.io/projected/e6ad8010-4207-4d04-9765-5b81facf9cad-kube-api-access-dmwhc\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.536137 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-scripts\") pod \"e6ad8010-4207-4d04-9765-5b81facf9cad\" (UID: \"e6ad8010-4207-4d04-9765-5b81facf9cad\") " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.537360 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-logs" (OuterVolumeSpecName: "logs") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.543970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.564568 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-scripts" (OuterVolumeSpecName: "scripts") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.573215 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.579702 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6ad8010-4207-4d04-9765-5b81facf9cad-kube-api-access-dmwhc" (OuterVolumeSpecName: "kube-api-access-dmwhc") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "kube-api-access-dmwhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.610309 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.639260 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.639445 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.639501 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.639581 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.639639 4758 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6ad8010-4207-4d04-9765-5b81facf9cad-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.639691 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmwhc\" (UniqueName: \"kubernetes.io/projected/e6ad8010-4207-4d04-9765-5b81facf9cad-kube-api-access-dmwhc\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.653092 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-config-data" (OuterVolumeSpecName: "config-data") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.665496 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.674828 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e6ad8010-4207-4d04-9765-5b81facf9cad" (UID: "e6ad8010-4207-4d04-9765-5b81facf9cad"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.740908 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.740939 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:03 crc kubenswrapper[4758]: I1004 11:09:03.740950 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6ad8010-4207-4d04-9765-5b81facf9cad-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.097398 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6ad8010-4207-4d04-9765-5b81facf9cad","Type":"ContainerDied","Data":"6f624be0f16857efaae6834175be97e7e417e994431be5bf0a463a54a2dfef5a"} Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.097454 4758 scope.go:117] "RemoveContainer" containerID="4ab23623bbd35983978197a60e4cbe69801b418dee40c090e20920c56ad37e18" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.097500 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.122601 4758 scope.go:117] "RemoveContainer" containerID="190b414813fbd25927e7ab3d30a73c8c1449cd4df4fb8e8945f3c6f5e3b9cbfb" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.140772 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.155472 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.164211 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:09:04 crc kubenswrapper[4758]: E1004 11:09:04.164530 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-log" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.164547 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-log" Oct 04 11:09:04 crc kubenswrapper[4758]: E1004 11:09:04.164584 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-httpd" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.164591 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-httpd" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.164736 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-httpd" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.164755 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" containerName="glance-log" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.165601 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.169577 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.169743 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.192713 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.248992 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69979b78-01c0-491b-8ab3-71aed97a8094-logs\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249070 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249187 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d658d\" (UniqueName: \"kubernetes.io/projected/69979b78-01c0-491b-8ab3-71aed97a8094-kube-api-access-d658d\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249209 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249284 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249410 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.249500 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69979b78-01c0-491b-8ab3-71aed97a8094-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.351431 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.351613 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352551 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d658d\" (UniqueName: \"kubernetes.io/projected/69979b78-01c0-491b-8ab3-71aed97a8094-kube-api-access-d658d\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352595 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352634 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352655 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352716 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352809 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69979b78-01c0-491b-8ab3-71aed97a8094-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.352876 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69979b78-01c0-491b-8ab3-71aed97a8094-logs\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.354646 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/69979b78-01c0-491b-8ab3-71aed97a8094-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.355023 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69979b78-01c0-491b-8ab3-71aed97a8094-logs\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.368201 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.368776 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.370765 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-config-data\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.377378 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/69979b78-01c0-491b-8ab3-71aed97a8094-scripts\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.377913 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d658d\" (UniqueName: \"kubernetes.io/projected/69979b78-01c0-491b-8ab3-71aed97a8094-kube-api-access-d658d\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.381526 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"69979b78-01c0-491b-8ab3-71aed97a8094\") " pod="openstack/glance-default-internal-api-0" Oct 04 11:09:04 crc kubenswrapper[4758]: I1004 11:09:04.485727 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.145501 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.155653 4758 generic.go:334] "Generic (PLEG): container finished" podID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerID="6f8f9aa39bf4e0f9918645575c022dce250e9ca8cf631068777b1e92c68b1fce" exitCode=0 Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.155713 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerDied","Data":"6f8f9aa39bf4e0f9918645575c022dce250e9ca8cf631068777b1e92c68b1fce"} Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.155738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d53f6f-4eee-4ea4-8e0a-20daac629083","Type":"ContainerDied","Data":"fd49949d2ebc0e613584f0607fae9fde577c3043d5c7836ad208c137112afffb"} Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.155747 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd49949d2ebc0e613584f0607fae9fde577c3043d5c7836ad208c137112afffb" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.241854 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270375 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx8xd\" (UniqueName: \"kubernetes.io/projected/92d53f6f-4eee-4ea4-8e0a-20daac629083-kube-api-access-nx8xd\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270470 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-combined-ca-bundle\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270498 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-scripts\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270528 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-run-httpd\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270548 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-config-data\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270568 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-log-httpd\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.270651 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-sg-core-conf-yaml\") pod \"92d53f6f-4eee-4ea4-8e0a-20daac629083\" (UID: \"92d53f6f-4eee-4ea4-8e0a-20daac629083\") " Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.271115 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.271358 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.276475 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d53f6f-4eee-4ea4-8e0a-20daac629083-kube-api-access-nx8xd" (OuterVolumeSpecName: "kube-api-access-nx8xd") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "kube-api-access-nx8xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.289563 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-scripts" (OuterVolumeSpecName: "scripts") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.300213 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.363271 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6ad8010-4207-4d04-9765-5b81facf9cad" path="/var/lib/kubelet/pods/e6ad8010-4207-4d04-9765-5b81facf9cad/volumes" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.372227 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx8xd\" (UniqueName: \"kubernetes.io/projected/92d53f6f-4eee-4ea4-8e0a-20daac629083-kube-api-access-nx8xd\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.372254 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.372263 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.372272 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d53f6f-4eee-4ea4-8e0a-20daac629083-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.372279 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.393069 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.415257 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-config-data" (OuterVolumeSpecName: "config-data") pod "92d53f6f-4eee-4ea4-8e0a-20daac629083" (UID: "92d53f6f-4eee-4ea4-8e0a-20daac629083"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.474213 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:05 crc kubenswrapper[4758]: I1004 11:09:05.474245 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d53f6f-4eee-4ea4-8e0a-20daac629083-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.169409 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.169590 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69979b78-01c0-491b-8ab3-71aed97a8094","Type":"ContainerStarted","Data":"b2657773edd0cf026f1ef4894496bbdb1c469b1f0176f63dd71a90537460311c"} Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.170221 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69979b78-01c0-491b-8ab3-71aed97a8094","Type":"ContainerStarted","Data":"71fa97353ddbfeb12e9e41986a13c4cdb880c1a5661308888970a89a8460f0e4"} Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.200080 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.209226 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.246183 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:06 crc kubenswrapper[4758]: E1004 11:09:06.246951 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="proxy-httpd" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.246975 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="proxy-httpd" Oct 04 11:09:06 crc kubenswrapper[4758]: E1004 11:09:06.247009 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="sg-core" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247015 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="sg-core" Oct 04 11:09:06 crc kubenswrapper[4758]: E1004 11:09:06.247042 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-notification-agent" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247049 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-notification-agent" Oct 04 11:09:06 crc kubenswrapper[4758]: E1004 11:09:06.247063 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-central-agent" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247070 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-central-agent" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247406 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="proxy-httpd" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247440 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-notification-agent" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247461 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="sg-core" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.247476 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" containerName="ceilometer-central-agent" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.297460 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.297565 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.303516 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.304276 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403174 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4ff7\" (UniqueName: \"kubernetes.io/projected/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-kube-api-access-p4ff7\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403221 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-config-data\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403293 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-run-httpd\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403316 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403337 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-log-httpd\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403391 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.403408 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-scripts\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.409087 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-94f8-account-create-7474t"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.410155 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.415566 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.427483 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-94f8-account-create-7474t"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505223 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-run-httpd\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505271 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505297 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-log-httpd\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505379 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505402 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-scripts\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505437 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4ff7\" (UniqueName: \"kubernetes.io/projected/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-kube-api-access-p4ff7\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505466 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-config-data\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505514 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lqv\" (UniqueName: \"kubernetes.io/projected/f503482b-df1b-4849-ad88-7a723462b9d4-kube-api-access-85lqv\") pod \"nova-api-94f8-account-create-7474t\" (UID: \"f503482b-df1b-4849-ad88-7a723462b9d4\") " pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.505995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-log-httpd\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.506194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-run-httpd\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.510482 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-config-data\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.511842 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.512384 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.514024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-scripts\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.531648 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4ff7\" (UniqueName: \"kubernetes.io/projected/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-kube-api-access-p4ff7\") pod \"ceilometer-0\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.607531 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lqv\" (UniqueName: \"kubernetes.io/projected/f503482b-df1b-4849-ad88-7a723462b9d4-kube-api-access-85lqv\") pod \"nova-api-94f8-account-create-7474t\" (UID: \"f503482b-df1b-4849-ad88-7a723462b9d4\") " pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.610076 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-1ee5-account-create-zfbg7"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.611386 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.612816 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.626210 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.635686 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1ee5-account-create-zfbg7"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.696011 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lqv\" (UniqueName: \"kubernetes.io/projected/f503482b-df1b-4849-ad88-7a723462b9d4-kube-api-access-85lqv\") pod \"nova-api-94f8-account-create-7474t\" (UID: \"f503482b-df1b-4849-ad88-7a723462b9d4\") " pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.711647 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb2h5\" (UniqueName: \"kubernetes.io/projected/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf-kube-api-access-nb2h5\") pod \"nova-cell0-1ee5-account-create-zfbg7\" (UID: \"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf\") " pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.727349 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.837014 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb2h5\" (UniqueName: \"kubernetes.io/projected/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf-kube-api-access-nb2h5\") pod \"nova-cell0-1ee5-account-create-zfbg7\" (UID: \"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf\") " pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.859392 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb2h5\" (UniqueName: \"kubernetes.io/projected/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf-kube-api-access-nb2h5\") pod \"nova-cell0-1ee5-account-create-zfbg7\" (UID: \"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf\") " pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.869075 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-18f9-account-create-46nw9"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.870546 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.873140 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.904261 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-18f9-account-create-46nw9"] Oct 04 11:09:06 crc kubenswrapper[4758]: I1004 11:09:06.928704 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.039934 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cwl7\" (UniqueName: \"kubernetes.io/projected/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2-kube-api-access-7cwl7\") pod \"nova-cell1-18f9-account-create-46nw9\" (UID: \"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2\") " pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.142320 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cwl7\" (UniqueName: \"kubernetes.io/projected/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2-kube-api-access-7cwl7\") pod \"nova-cell1-18f9-account-create-46nw9\" (UID: \"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2\") " pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.161622 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cwl7\" (UniqueName: \"kubernetes.io/projected/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2-kube-api-access-7cwl7\") pod \"nova-cell1-18f9-account-create-46nw9\" (UID: \"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2\") " pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.203274 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"69979b78-01c0-491b-8ab3-71aed97a8094","Type":"ContainerStarted","Data":"9e13f333a913e07fd2b2a3c6c6beae90e9aeb2e71eebfb76f3bdfdba6992df67"} Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.215611 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.234397 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.234374626 podStartE2EDuration="3.234374626s" podCreationTimestamp="2025-10-04 11:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:07.224006366 +0000 UTC m=+1124.516657255" watchObservedRunningTime="2025-10-04 11:09:07.234374626 +0000 UTC m=+1124.527025515" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.279697 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.342870 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d53f6f-4eee-4ea4-8e0a-20daac629083" path="/var/lib/kubelet/pods/92d53f6f-4eee-4ea4-8e0a-20daac629083/volumes" Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.445525 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-94f8-account-create-7474t"] Oct 04 11:09:07 crc kubenswrapper[4758]: W1004 11:09:07.453016 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf503482b_df1b_4849_ad88_7a723462b9d4.slice/crio-510a0dbdb5b6710928a8ee96decb32d4f5a49f4d0b72ee4d8ca60501675fceb1 WatchSource:0}: Error finding container 510a0dbdb5b6710928a8ee96decb32d4f5a49f4d0b72ee4d8ca60501675fceb1: Status 404 returned error can't find the container with id 510a0dbdb5b6710928a8ee96decb32d4f5a49f4d0b72ee4d8ca60501675fceb1 Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.606896 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1ee5-account-create-zfbg7"] Oct 04 11:09:07 crc kubenswrapper[4758]: W1004 11:09:07.617281 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd746b6d_78c5_4ef6_a466_ad7ed7d346cf.slice/crio-169f0d34291037131b59d80598b67d53b1f0b190071eaa294eec4987dded3b1d WatchSource:0}: Error finding container 169f0d34291037131b59d80598b67d53b1f0b190071eaa294eec4987dded3b1d: Status 404 returned error can't find the container with id 169f0d34291037131b59d80598b67d53b1f0b190071eaa294eec4987dded3b1d Oct 04 11:09:07 crc kubenswrapper[4758]: I1004 11:09:07.741946 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-18f9-account-create-46nw9"] Oct 04 11:09:07 crc kubenswrapper[4758]: W1004 11:09:07.746429 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a17d34_49c8_4ad5_8f2d_86b9b665dfa2.slice/crio-37b90dedeaa7a9ed305f913d4c9597d871a0362e73183ce635778778caa4cd37 WatchSource:0}: Error finding container 37b90dedeaa7a9ed305f913d4c9597d871a0362e73183ce635778778caa4cd37: Status 404 returned error can't find the container with id 37b90dedeaa7a9ed305f913d4c9597d871a0362e73183ce635778778caa4cd37 Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.213846 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" event={"ID":"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf","Type":"ContainerStarted","Data":"ffe8fedcfe68f06d92fb98d31f6c1b7f90a7ccb82d7b5662cd6ce39b8cd39fc2"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.213887 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" event={"ID":"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf","Type":"ContainerStarted","Data":"169f0d34291037131b59d80598b67d53b1f0b190071eaa294eec4987dded3b1d"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.215528 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerStarted","Data":"d90911d2bbdc156475eec402aa767b05b1c9ff27e2f55b5d22e3e57dc96c9604"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.218028 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-94f8-account-create-7474t" event={"ID":"f503482b-df1b-4849-ad88-7a723462b9d4","Type":"ContainerStarted","Data":"fc928abe78917a701c13580ece3f66592a29df27fe9f44b82ae173b257641ef9"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.218239 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-94f8-account-create-7474t" event={"ID":"f503482b-df1b-4849-ad88-7a723462b9d4","Type":"ContainerStarted","Data":"510a0dbdb5b6710928a8ee96decb32d4f5a49f4d0b72ee4d8ca60501675fceb1"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.225979 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-18f9-account-create-46nw9" event={"ID":"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2","Type":"ContainerStarted","Data":"d9a6f7fe3bc5b63061bb8c05a6529699471f771b3f5063ad38e7d9af73ebd0f3"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.226024 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-18f9-account-create-46nw9" event={"ID":"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2","Type":"ContainerStarted","Data":"37b90dedeaa7a9ed305f913d4c9597d871a0362e73183ce635778778caa4cd37"} Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.242333 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" podStartSLOduration=2.242290738 podStartE2EDuration="2.242290738s" podCreationTimestamp="2025-10-04 11:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:08.231368623 +0000 UTC m=+1125.524019512" watchObservedRunningTime="2025-10-04 11:09:08.242290738 +0000 UTC m=+1125.534941627" Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.246513 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-94f8-account-create-7474t" podStartSLOduration=2.246500612 podStartE2EDuration="2.246500612s" podCreationTimestamp="2025-10-04 11:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:08.241842336 +0000 UTC m=+1125.534493225" watchObservedRunningTime="2025-10-04 11:09:08.246500612 +0000 UTC m=+1125.539151501" Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.262148 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-18f9-account-create-46nw9" podStartSLOduration=2.262126783 podStartE2EDuration="2.262126783s" podCreationTimestamp="2025-10-04 11:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:08.254252291 +0000 UTC m=+1125.546903200" watchObservedRunningTime="2025-10-04 11:09:08.262126783 +0000 UTC m=+1125.554777682" Oct 04 11:09:08 crc kubenswrapper[4758]: I1004 11:09:08.755580 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c98cb656b-lczgq" podUID="61f721a8-00fb-446c-8e98-922827d2fc39" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.236310 4758 generic.go:334] "Generic (PLEG): container finished" podID="cd746b6d-78c5-4ef6-a466-ad7ed7d346cf" containerID="ffe8fedcfe68f06d92fb98d31f6c1b7f90a7ccb82d7b5662cd6ce39b8cd39fc2" exitCode=0 Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.236678 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" event={"ID":"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf","Type":"ContainerDied","Data":"ffe8fedcfe68f06d92fb98d31f6c1b7f90a7ccb82d7b5662cd6ce39b8cd39fc2"} Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.238080 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerStarted","Data":"32699b8996125709f01b8830b8e0006ecd2623f3e9fe8579e90de22143e9e4a6"} Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.239981 4758 generic.go:334] "Generic (PLEG): container finished" podID="f503482b-df1b-4849-ad88-7a723462b9d4" containerID="fc928abe78917a701c13580ece3f66592a29df27fe9f44b82ae173b257641ef9" exitCode=0 Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.240066 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-94f8-account-create-7474t" event={"ID":"f503482b-df1b-4849-ad88-7a723462b9d4","Type":"ContainerDied","Data":"fc928abe78917a701c13580ece3f66592a29df27fe9f44b82ae173b257641ef9"} Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.241965 4758 generic.go:334] "Generic (PLEG): container finished" podID="51a17d34-49c8-4ad5-8f2d-86b9b665dfa2" containerID="d9a6f7fe3bc5b63061bb8c05a6529699471f771b3f5063ad38e7d9af73ebd0f3" exitCode=0 Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.241999 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-18f9-account-create-46nw9" event={"ID":"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2","Type":"ContainerDied","Data":"d9a6f7fe3bc5b63061bb8c05a6529699471f771b3f5063ad38e7d9af73ebd0f3"} Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.449411 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.449768 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.504703 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 11:09:09 crc kubenswrapper[4758]: I1004 11:09:09.508065 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.065859 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.251440 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerStarted","Data":"e625d603b36048d4d626c968262e5990733651b4324e0505230271c07637d817"} Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.251826 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.251845 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.733774 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.831209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb2h5\" (UniqueName: \"kubernetes.io/projected/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf-kube-api-access-nb2h5\") pod \"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf\" (UID: \"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf\") " Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.839593 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf-kube-api-access-nb2h5" (OuterVolumeSpecName: "kube-api-access-nb2h5") pod "cd746b6d-78c5-4ef6-a466-ad7ed7d346cf" (UID: "cd746b6d-78c5-4ef6-a466-ad7ed7d346cf"). InnerVolumeSpecName "kube-api-access-nb2h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.932893 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb2h5\" (UniqueName: \"kubernetes.io/projected/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf-kube-api-access-nb2h5\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:10 crc kubenswrapper[4758]: I1004 11:09:10.992201 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.010600 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.134834 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85lqv\" (UniqueName: \"kubernetes.io/projected/f503482b-df1b-4849-ad88-7a723462b9d4-kube-api-access-85lqv\") pod \"f503482b-df1b-4849-ad88-7a723462b9d4\" (UID: \"f503482b-df1b-4849-ad88-7a723462b9d4\") " Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.135200 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cwl7\" (UniqueName: \"kubernetes.io/projected/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2-kube-api-access-7cwl7\") pod \"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2\" (UID: \"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2\") " Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.149387 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2-kube-api-access-7cwl7" (OuterVolumeSpecName: "kube-api-access-7cwl7") pod "51a17d34-49c8-4ad5-8f2d-86b9b665dfa2" (UID: "51a17d34-49c8-4ad5-8f2d-86b9b665dfa2"). InnerVolumeSpecName "kube-api-access-7cwl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.153303 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f503482b-df1b-4849-ad88-7a723462b9d4-kube-api-access-85lqv" (OuterVolumeSpecName: "kube-api-access-85lqv") pod "f503482b-df1b-4849-ad88-7a723462b9d4" (UID: "f503482b-df1b-4849-ad88-7a723462b9d4"). InnerVolumeSpecName "kube-api-access-85lqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.237424 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85lqv\" (UniqueName: \"kubernetes.io/projected/f503482b-df1b-4849-ad88-7a723462b9d4-kube-api-access-85lqv\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.237456 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cwl7\" (UniqueName: \"kubernetes.io/projected/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2-kube-api-access-7cwl7\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.260125 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-18f9-account-create-46nw9" event={"ID":"51a17d34-49c8-4ad5-8f2d-86b9b665dfa2","Type":"ContainerDied","Data":"37b90dedeaa7a9ed305f913d4c9597d871a0362e73183ce635778778caa4cd37"} Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.260160 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b90dedeaa7a9ed305f913d4c9597d871a0362e73183ce635778778caa4cd37" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.260205 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-18f9-account-create-46nw9" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.279224 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" event={"ID":"cd746b6d-78c5-4ef6-a466-ad7ed7d346cf","Type":"ContainerDied","Data":"169f0d34291037131b59d80598b67d53b1f0b190071eaa294eec4987dded3b1d"} Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.279389 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="169f0d34291037131b59d80598b67d53b1f0b190071eaa294eec4987dded3b1d" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.279504 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1ee5-account-create-zfbg7" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.282260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerStarted","Data":"3ef31504eeff9d9313a82f5a7dd8c057eb7c926df1d667090e0d90ae7c653d0e"} Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.285450 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-94f8-account-create-7474t" Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.285722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-94f8-account-create-7474t" event={"ID":"f503482b-df1b-4849-ad88-7a723462b9d4","Type":"ContainerDied","Data":"510a0dbdb5b6710928a8ee96decb32d4f5a49f4d0b72ee4d8ca60501675fceb1"} Oct 04 11:09:11 crc kubenswrapper[4758]: I1004 11:09:11.285773 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="510a0dbdb5b6710928a8ee96decb32d4f5a49f4d0b72ee4d8ca60501675fceb1" Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.086440 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.087122 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.094155 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.304249 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerStarted","Data":"060278966d9312eed1a01e30af470add7e2d7e2c7a0966360046f871906b47c1"} Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.304689 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-central-agent" containerID="cri-o://32699b8996125709f01b8830b8e0006ecd2623f3e9fe8579e90de22143e9e4a6" gracePeriod=30 Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.305051 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="proxy-httpd" containerID="cri-o://060278966d9312eed1a01e30af470add7e2d7e2c7a0966360046f871906b47c1" gracePeriod=30 Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.305095 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-notification-agent" containerID="cri-o://e625d603b36048d4d626c968262e5990733651b4324e0505230271c07637d817" gracePeriod=30 Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.305188 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="sg-core" containerID="cri-o://3ef31504eeff9d9313a82f5a7dd8c057eb7c926df1d667090e0d90ae7c653d0e" gracePeriod=30 Oct 04 11:09:13 crc kubenswrapper[4758]: I1004 11:09:13.341336 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.211932877 podStartE2EDuration="7.34131933s" podCreationTimestamp="2025-10-04 11:09:06 +0000 UTC" firstStartedPulling="2025-10-04 11:09:07.290724157 +0000 UTC m=+1124.583375046" lastFinishedPulling="2025-10-04 11:09:12.42011061 +0000 UTC m=+1129.712761499" observedRunningTime="2025-10-04 11:09:13.331941737 +0000 UTC m=+1130.624592626" watchObservedRunningTime="2025-10-04 11:09:13.34131933 +0000 UTC m=+1130.633970219" Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.314409 4758 generic.go:334] "Generic (PLEG): container finished" podID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerID="060278966d9312eed1a01e30af470add7e2d7e2c7a0966360046f871906b47c1" exitCode=0 Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.315509 4758 generic.go:334] "Generic (PLEG): container finished" podID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerID="3ef31504eeff9d9313a82f5a7dd8c057eb7c926df1d667090e0d90ae7c653d0e" exitCode=2 Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.315595 4758 generic.go:334] "Generic (PLEG): container finished" podID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerID="e625d603b36048d4d626c968262e5990733651b4324e0505230271c07637d817" exitCode=0 Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.315664 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerDied","Data":"060278966d9312eed1a01e30af470add7e2d7e2c7a0966360046f871906b47c1"} Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.315736 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerDied","Data":"3ef31504eeff9d9313a82f5a7dd8c057eb7c926df1d667090e0d90ae7c653d0e"} Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.315792 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerDied","Data":"e625d603b36048d4d626c968262e5990733651b4324e0505230271c07637d817"} Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.486888 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.491022 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.519653 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:14 crc kubenswrapper[4758]: I1004 11:09:14.563567 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:15 crc kubenswrapper[4758]: I1004 11:09:15.336440 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:15 crc kubenswrapper[4758]: I1004 11:09:15.336482 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.360683 4758 generic.go:334] "Generic (PLEG): container finished" podID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerID="32699b8996125709f01b8830b8e0006ecd2623f3e9fe8579e90de22143e9e4a6" exitCode=0 Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.361857 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerDied","Data":"32699b8996125709f01b8830b8e0006ecd2623f3e9fe8579e90de22143e9e4a6"} Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.602584 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732288 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-run-httpd\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732745 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-combined-ca-bundle\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732820 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-log-httpd\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732875 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-sg-core-conf-yaml\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732909 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4ff7\" (UniqueName: \"kubernetes.io/projected/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-kube-api-access-p4ff7\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732959 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-scripts\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.732994 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-config-data\") pod \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\" (UID: \"e50afcd2-84d2-4fc5-9509-8112c4dffd1e\") " Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.734763 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.735970 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.743254 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-kube-api-access-p4ff7" (OuterVolumeSpecName: "kube-api-access-p4ff7") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "kube-api-access-p4ff7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.744244 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-scripts" (OuterVolumeSpecName: "scripts") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.768243 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.836185 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.836219 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.836229 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.836239 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4ff7\" (UniqueName: \"kubernetes.io/projected/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-kube-api-access-p4ff7\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.836248 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.893746 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.941800 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:16 crc kubenswrapper[4758]: I1004 11:09:16.974277 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-config-data" (OuterVolumeSpecName: "config-data") pod "e50afcd2-84d2-4fc5-9509-8112c4dffd1e" (UID: "e50afcd2-84d2-4fc5-9509-8112c4dffd1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.043338 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e50afcd2-84d2-4fc5-9509-8112c4dffd1e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167125 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8tq5p"] Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167536 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-central-agent" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167562 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-central-agent" Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167583 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-notification-agent" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167591 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-notification-agent" Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167604 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="proxy-httpd" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167613 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="proxy-httpd" Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167648 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd746b6d-78c5-4ef6-a466-ad7ed7d346cf" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167656 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd746b6d-78c5-4ef6-a466-ad7ed7d346cf" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167663 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="sg-core" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167671 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="sg-core" Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167689 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f503482b-df1b-4849-ad88-7a723462b9d4" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167696 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f503482b-df1b-4849-ad88-7a723462b9d4" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: E1004 11:09:17.167709 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a17d34-49c8-4ad5-8f2d-86b9b665dfa2" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.167718 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a17d34-49c8-4ad5-8f2d-86b9b665dfa2" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168069 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f503482b-df1b-4849-ad88-7a723462b9d4" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168090 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a17d34-49c8-4ad5-8f2d-86b9b665dfa2" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168121 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="sg-core" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168132 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-central-agent" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168147 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="ceilometer-notification-agent" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168162 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd746b6d-78c5-4ef6-a466-ad7ed7d346cf" containerName="mariadb-account-create" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168181 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" containerName="proxy-httpd" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.168874 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.172199 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.172447 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hcfdn" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.172595 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.183388 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8tq5p"] Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.348161 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-scripts\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.348232 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-config-data\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.348285 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.348314 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6lnt\" (UniqueName: \"kubernetes.io/projected/56103956-1715-4a93-9784-70dd825c43a2-kube-api-access-q6lnt\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.371995 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e50afcd2-84d2-4fc5-9509-8112c4dffd1e","Type":"ContainerDied","Data":"d90911d2bbdc156475eec402aa767b05b1c9ff27e2f55b5d22e3e57dc96c9604"} Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.372048 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.372013 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.372086 4758 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.372066 4758 scope.go:117] "RemoveContainer" containerID="060278966d9312eed1a01e30af470add7e2d7e2c7a0966360046f871906b47c1" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.392329 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.392484 4758 scope.go:117] "RemoveContainer" containerID="3ef31504eeff9d9313a82f5a7dd8c057eb7c926df1d667090e0d90ae7c653d0e" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.406360 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.413658 4758 scope.go:117] "RemoveContainer" containerID="e625d603b36048d4d626c968262e5990733651b4324e0505230271c07637d817" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.417619 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.420025 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.424973 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.425653 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.444172 4758 scope.go:117] "RemoveContainer" containerID="32699b8996125709f01b8830b8e0006ecd2623f3e9fe8579e90de22143e9e4a6" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.449993 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-scripts\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.450072 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-config-data\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.450161 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.450192 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6lnt\" (UniqueName: \"kubernetes.io/projected/56103956-1715-4a93-9784-70dd825c43a2-kube-api-access-q6lnt\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.454943 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.455610 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-scripts\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.458014 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.459680 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-config-data\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.470340 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6lnt\" (UniqueName: \"kubernetes.io/projected/56103956-1715-4a93-9784-70dd825c43a2-kube-api-access-q6lnt\") pod \"nova-cell0-conductor-db-sync-8tq5p\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.514626 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551463 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551631 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-run-httpd\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551663 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-scripts\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551701 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551724 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-config-data\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551752 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-log-httpd\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.551770 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mwzk\" (UniqueName: \"kubernetes.io/projected/3520d415-f35c-44f0-a001-638478475b44-kube-api-access-2mwzk\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.653610 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-run-httpd\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.653939 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-scripts\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.653978 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.653996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-config-data\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.654017 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-log-httpd\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.654036 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mwzk\" (UniqueName: \"kubernetes.io/projected/3520d415-f35c-44f0-a001-638478475b44-kube-api-access-2mwzk\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.654076 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.654634 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-run-httpd\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.655497 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-log-httpd\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.662274 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-config-data\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.662993 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.664356 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.665304 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-scripts\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.671421 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mwzk\" (UniqueName: \"kubernetes.io/projected/3520d415-f35c-44f0-a001-638478475b44-kube-api-access-2mwzk\") pod \"ceilometer-0\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.718078 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.730611 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.745749 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:09:17 crc kubenswrapper[4758]: I1004 11:09:17.836674 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8tq5p"] Oct 04 11:09:18 crc kubenswrapper[4758]: I1004 11:09:18.342701 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:18 crc kubenswrapper[4758]: I1004 11:09:18.381119 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" event={"ID":"56103956-1715-4a93-9784-70dd825c43a2","Type":"ContainerStarted","Data":"dfa24a27b0b087d83b419e3bf8c2adf7ddabb0f041ba2591bbde1635e5d30d81"} Oct 04 11:09:18 crc kubenswrapper[4758]: I1004 11:09:18.384300 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerStarted","Data":"9666266804d214cc078e4f2cba765d4789c776b0b390311394a8c653a65f0cbb"} Oct 04 11:09:18 crc kubenswrapper[4758]: I1004 11:09:18.756854 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c98cb656b-lczgq" podUID="61f721a8-00fb-446c-8e98-922827d2fc39" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 11:09:19 crc kubenswrapper[4758]: I1004 11:09:19.336327 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50afcd2-84d2-4fc5-9509-8112c4dffd1e" path="/var/lib/kubelet/pods/e50afcd2-84d2-4fc5-9509-8112c4dffd1e/volumes" Oct 04 11:09:19 crc kubenswrapper[4758]: I1004 11:09:19.396624 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerStarted","Data":"84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483"} Oct 04 11:09:20 crc kubenswrapper[4758]: I1004 11:09:20.414314 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerStarted","Data":"43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6"} Oct 04 11:09:28 crc kubenswrapper[4758]: I1004 11:09:28.497586 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" event={"ID":"56103956-1715-4a93-9784-70dd825c43a2","Type":"ContainerStarted","Data":"9ffc9eaa9ef2818852dcbdd662ca88e1126b420be8d0d631c7afdc3d45f32d27"} Oct 04 11:09:28 crc kubenswrapper[4758]: I1004 11:09:28.500128 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerStarted","Data":"b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56"} Oct 04 11:09:28 crc kubenswrapper[4758]: I1004 11:09:28.516665 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" podStartSLOduration=1.991189083 podStartE2EDuration="11.516649299s" podCreationTimestamp="2025-10-04 11:09:17 +0000 UTC" firstStartedPulling="2025-10-04 11:09:17.915078301 +0000 UTC m=+1135.207729190" lastFinishedPulling="2025-10-04 11:09:27.440538517 +0000 UTC m=+1144.733189406" observedRunningTime="2025-10-04 11:09:28.510051831 +0000 UTC m=+1145.802702720" watchObservedRunningTime="2025-10-04 11:09:28.516649299 +0000 UTC m=+1145.809300188" Oct 04 11:09:29 crc kubenswrapper[4758]: I1004 11:09:29.510485 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerStarted","Data":"eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099"} Oct 04 11:09:29 crc kubenswrapper[4758]: I1004 11:09:29.511497 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 11:09:29 crc kubenswrapper[4758]: I1004 11:09:29.539872 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.044314279 podStartE2EDuration="12.539850874s" podCreationTimestamp="2025-10-04 11:09:17 +0000 UTC" firstStartedPulling="2025-10-04 11:09:18.349191642 +0000 UTC m=+1135.641842531" lastFinishedPulling="2025-10-04 11:09:28.844728237 +0000 UTC m=+1146.137379126" observedRunningTime="2025-10-04 11:09:29.531990192 +0000 UTC m=+1146.824641081" watchObservedRunningTime="2025-10-04 11:09:29.539850874 +0000 UTC m=+1146.832501763" Oct 04 11:09:30 crc kubenswrapper[4758]: I1004 11:09:30.717379 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:09:32 crc kubenswrapper[4758]: I1004 11:09:32.587658 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5c98cb656b-lczgq" Oct 04 11:09:32 crc kubenswrapper[4758]: I1004 11:09:32.707399 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-549f9cdcb8-66q22"] Oct 04 11:09:32 crc kubenswrapper[4758]: I1004 11:09:32.708927 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon-log" containerID="cri-o://e10f41f31bbb4f5fd1672270aaecda9dd129769e4c72c63717d0592b51d67711" gracePeriod=30 Oct 04 11:09:32 crc kubenswrapper[4758]: I1004 11:09:32.709156 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" containerID="cri-o://b7c2b7c398b965f1af57612a5dd546394a6cac04c79b30a09d3ac182e4e98488" gracePeriod=30 Oct 04 11:09:36 crc kubenswrapper[4758]: I1004 11:09:36.571649 4758 generic.go:334] "Generic (PLEG): container finished" podID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerID="b7c2b7c398b965f1af57612a5dd546394a6cac04c79b30a09d3ac182e4e98488" exitCode=0 Oct 04 11:09:36 crc kubenswrapper[4758]: I1004 11:09:36.571727 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549f9cdcb8-66q22" event={"ID":"55a162df-bf26-4c9f-9808-506db4c3bd01","Type":"ContainerDied","Data":"b7c2b7c398b965f1af57612a5dd546394a6cac04c79b30a09d3ac182e4e98488"} Oct 04 11:09:38 crc kubenswrapper[4758]: I1004 11:09:38.574095 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 04 11:09:40 crc kubenswrapper[4758]: I1004 11:09:40.623996 4758 generic.go:334] "Generic (PLEG): container finished" podID="56103956-1715-4a93-9784-70dd825c43a2" containerID="9ffc9eaa9ef2818852dcbdd662ca88e1126b420be8d0d631c7afdc3d45f32d27" exitCode=0 Oct 04 11:09:40 crc kubenswrapper[4758]: I1004 11:09:40.625524 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" event={"ID":"56103956-1715-4a93-9784-70dd825c43a2","Type":"ContainerDied","Data":"9ffc9eaa9ef2818852dcbdd662ca88e1126b420be8d0d631c7afdc3d45f32d27"} Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.023398 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.092552 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6lnt\" (UniqueName: \"kubernetes.io/projected/56103956-1715-4a93-9784-70dd825c43a2-kube-api-access-q6lnt\") pod \"56103956-1715-4a93-9784-70dd825c43a2\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.092683 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-combined-ca-bundle\") pod \"56103956-1715-4a93-9784-70dd825c43a2\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.092734 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-scripts\") pod \"56103956-1715-4a93-9784-70dd825c43a2\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.092779 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-config-data\") pod \"56103956-1715-4a93-9784-70dd825c43a2\" (UID: \"56103956-1715-4a93-9784-70dd825c43a2\") " Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.101774 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-scripts" (OuterVolumeSpecName: "scripts") pod "56103956-1715-4a93-9784-70dd825c43a2" (UID: "56103956-1715-4a93-9784-70dd825c43a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.102143 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56103956-1715-4a93-9784-70dd825c43a2-kube-api-access-q6lnt" (OuterVolumeSpecName: "kube-api-access-q6lnt") pod "56103956-1715-4a93-9784-70dd825c43a2" (UID: "56103956-1715-4a93-9784-70dd825c43a2"). InnerVolumeSpecName "kube-api-access-q6lnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.128198 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-config-data" (OuterVolumeSpecName: "config-data") pod "56103956-1715-4a93-9784-70dd825c43a2" (UID: "56103956-1715-4a93-9784-70dd825c43a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.139283 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56103956-1715-4a93-9784-70dd825c43a2" (UID: "56103956-1715-4a93-9784-70dd825c43a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.195075 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6lnt\" (UniqueName: \"kubernetes.io/projected/56103956-1715-4a93-9784-70dd825c43a2-kube-api-access-q6lnt\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.195126 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.195139 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.195147 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56103956-1715-4a93-9784-70dd825c43a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.646334 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" event={"ID":"56103956-1715-4a93-9784-70dd825c43a2","Type":"ContainerDied","Data":"dfa24a27b0b087d83b419e3bf8c2adf7ddabb0f041ba2591bbde1635e5d30d81"} Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.646822 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfa24a27b0b087d83b419e3bf8c2adf7ddabb0f041ba2591bbde1635e5d30d81" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.646393 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8tq5p" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.804609 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 11:09:42 crc kubenswrapper[4758]: E1004 11:09:42.804977 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56103956-1715-4a93-9784-70dd825c43a2" containerName="nova-cell0-conductor-db-sync" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.804995 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="56103956-1715-4a93-9784-70dd825c43a2" containerName="nova-cell0-conductor-db-sync" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.805164 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="56103956-1715-4a93-9784-70dd825c43a2" containerName="nova-cell0-conductor-db-sync" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.805749 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.808185 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.808726 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hcfdn" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.829695 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.916844 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c88fa4e-31fc-4857-9934-e63c956f00f0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.916893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c88fa4e-31fc-4857-9934-e63c956f00f0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:42 crc kubenswrapper[4758]: I1004 11:09:42.916923 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chpvt\" (UniqueName: \"kubernetes.io/projected/4c88fa4e-31fc-4857-9934-e63c956f00f0-kube-api-access-chpvt\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.018780 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c88fa4e-31fc-4857-9934-e63c956f00f0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.018836 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c88fa4e-31fc-4857-9934-e63c956f00f0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.018864 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chpvt\" (UniqueName: \"kubernetes.io/projected/4c88fa4e-31fc-4857-9934-e63c956f00f0-kube-api-access-chpvt\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.023512 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c88fa4e-31fc-4857-9934-e63c956f00f0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.025476 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c88fa4e-31fc-4857-9934-e63c956f00f0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.042452 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chpvt\" (UniqueName: \"kubernetes.io/projected/4c88fa4e-31fc-4857-9934-e63c956f00f0-kube-api-access-chpvt\") pod \"nova-cell0-conductor-0\" (UID: \"4c88fa4e-31fc-4857-9934-e63c956f00f0\") " pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.130350 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.628166 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 11:09:43 crc kubenswrapper[4758]: I1004 11:09:43.655203 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4c88fa4e-31fc-4857-9934-e63c956f00f0","Type":"ContainerStarted","Data":"e197c3d71fb8eeffc515e7623d8d0f62cea10eaa6b25e4e4406efb56af787bdb"} Oct 04 11:09:44 crc kubenswrapper[4758]: I1004 11:09:44.667573 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4c88fa4e-31fc-4857-9934-e63c956f00f0","Type":"ContainerStarted","Data":"99a77f12de49c0e07ee34ae5299fd27614aebfff1cf05e9985c6cb3184155316"} Oct 04 11:09:44 crc kubenswrapper[4758]: I1004 11:09:44.668087 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:44 crc kubenswrapper[4758]: I1004 11:09:44.686642 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.686622313 podStartE2EDuration="2.686622313s" podCreationTimestamp="2025-10-04 11:09:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:44.683853168 +0000 UTC m=+1161.976504057" watchObservedRunningTime="2025-10-04 11:09:44.686622313 +0000 UTC m=+1161.979273212" Oct 04 11:09:47 crc kubenswrapper[4758]: I1004 11:09:47.761421 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.155992 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.574775 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.757650 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-4p4g9"] Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.759166 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.761205 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.761765 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.768891 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4p4g9"] Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.836202 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-config-data\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.836300 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.836469 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zk66\" (UniqueName: \"kubernetes.io/projected/702dad08-3e1d-4118-9ecd-6c1882b8188d-kube-api-access-2zk66\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.836529 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-scripts\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.938606 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-scripts\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.938677 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-config-data\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.938731 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.938851 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zk66\" (UniqueName: \"kubernetes.io/projected/702dad08-3e1d-4118-9ecd-6c1882b8188d-kube-api-access-2zk66\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.970525 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-config-data\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.972579 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.990339 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-scripts\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:48 crc kubenswrapper[4758]: I1004 11:09:48.997236 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:48.998928 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.009530 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.018607 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zk66\" (UniqueName: \"kubernetes.io/projected/702dad08-3e1d-4118-9ecd-6c1882b8188d-kube-api-access-2zk66\") pod \"nova-cell0-cell-mapping-4p4g9\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.036270 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.040023 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-logs\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.040076 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.040177 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxgr2\" (UniqueName: \"kubernetes.io/projected/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-kube-api-access-pxgr2\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.040218 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-config-data\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.085128 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.144045 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.144611 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxgr2\" (UniqueName: \"kubernetes.io/projected/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-kube-api-access-pxgr2\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.144737 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-config-data\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.144839 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-logs\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.145252 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-logs\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.155827 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.160068 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.160227 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-config-data\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.170651 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.180950 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.205152 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-c4qcg"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.206587 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.214436 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxgr2\" (UniqueName: \"kubernetes.io/projected/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-kube-api-access-pxgr2\") pod \"nova-metadata-0\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.224303 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.267452 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.267537 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2wss\" (UniqueName: \"kubernetes.io/projected/56db99fe-c181-450e-b9cc-f8810b6658e8-kube-api-access-v2wss\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.267568 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-config\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.267613 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.279788 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-c4qcg"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.294326 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56gbv\" (UniqueName: \"kubernetes.io/projected/c710639c-0dc3-43ba-b084-fee7da532dc8-kube-api-access-56gbv\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.294410 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.294524 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-config-data\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.294573 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-svc\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.294606 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.381140 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.382535 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.399497 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.399739 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2wss\" (UniqueName: \"kubernetes.io/projected/56db99fe-c181-450e-b9cc-f8810b6658e8-kube-api-access-v2wss\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.399864 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-config\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.399967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.400052 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56gbv\" (UniqueName: \"kubernetes.io/projected/c710639c-0dc3-43ba-b084-fee7da532dc8-kube-api-access-56gbv\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.400152 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.401064 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-config-data\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.401247 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-svc\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.401324 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.401553 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.402051 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.402642 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-svc\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.400983 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-config\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.403786 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.406563 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.407063 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.417470 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.427244 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.438740 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-config-data\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.451051 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2wss\" (UniqueName: \"kubernetes.io/projected/56db99fe-c181-450e-b9cc-f8810b6658e8-kube-api-access-v2wss\") pod \"nova-scheduler-0\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.452638 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56gbv\" (UniqueName: \"kubernetes.io/projected/c710639c-0dc3-43ba-b084-fee7da532dc8-kube-api-access-56gbv\") pod \"dnsmasq-dns-757b4f8459-c4qcg\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.502713 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ee35768-f230-4b06-991a-31b79d2d3d78-logs\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.502889 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.502916 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxdpw\" (UniqueName: \"kubernetes.io/projected/0ee35768-f230-4b06-991a-31b79d2d3d78-kube-api-access-jxdpw\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.502936 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-config-data\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.523148 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.529555 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.531956 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.542355 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.577590 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.589475 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.604963 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.605018 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxdpw\" (UniqueName: \"kubernetes.io/projected/0ee35768-f230-4b06-991a-31b79d2d3d78-kube-api-access-jxdpw\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.605042 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-config-data\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.605130 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.605632 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.605667 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ee35768-f230-4b06-991a-31b79d2d3d78-logs\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.605776 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7gmm\" (UniqueName: \"kubernetes.io/projected/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-kube-api-access-h7gmm\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.606019 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ee35768-f230-4b06-991a-31b79d2d3d78-logs\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.611540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.611652 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-config-data\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.625522 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxdpw\" (UniqueName: \"kubernetes.io/projected/0ee35768-f230-4b06-991a-31b79d2d3d78-kube-api-access-jxdpw\") pod \"nova-api-0\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.708592 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7gmm\" (UniqueName: \"kubernetes.io/projected/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-kube-api-access-h7gmm\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.708664 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.708698 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.716762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.721609 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.724442 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7gmm\" (UniqueName: \"kubernetes.io/projected/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-kube-api-access-h7gmm\") pod \"nova-cell1-novncproxy-0\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.773753 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.852114 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:09:49 crc kubenswrapper[4758]: I1004 11:09:49.859409 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4p4g9"] Oct 04 11:09:49 crc kubenswrapper[4758]: W1004 11:09:49.866329 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod702dad08_3e1d_4118_9ecd_6c1882b8188d.slice/crio-dcdaa69f2a841e657b7250ac0fe687d3af1d4025ecfe411acb68695eb9ff6668 WatchSource:0}: Error finding container dcdaa69f2a841e657b7250ac0fe687d3af1d4025ecfe411acb68695eb9ff6668: Status 404 returned error can't find the container with id dcdaa69f2a841e657b7250ac0fe687d3af1d4025ecfe411acb68695eb9ff6668 Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.043788 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.178801 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-c4qcg"] Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.304979 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:09:50 crc kubenswrapper[4758]: W1004 11:09:50.352505 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56db99fe_c181_450e_b9cc_f8810b6658e8.slice/crio-da7d7d9adaab0130fe20d0fbe71603f9b1a8babebf020cc0aea96704eb53f5a3 WatchSource:0}: Error finding container da7d7d9adaab0130fe20d0fbe71603f9b1a8babebf020cc0aea96704eb53f5a3: Status 404 returned error can't find the container with id da7d7d9adaab0130fe20d0fbe71603f9b1a8babebf020cc0aea96704eb53f5a3 Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.376879 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gn24j"] Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.378625 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.385513 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.385867 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.393296 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gn24j"] Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.530816 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.531443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6zcx\" (UniqueName: \"kubernetes.io/projected/ac470595-559d-4607-9113-450378bcaf13-kube-api-access-d6zcx\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.531651 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-config-data\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.531768 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-scripts\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.531843 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.633751 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-config-data\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.633809 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-scripts\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.633842 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.633902 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6zcx\" (UniqueName: \"kubernetes.io/projected/ac470595-559d-4607-9113-450378bcaf13-kube-api-access-d6zcx\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.638187 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-scripts\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.638194 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.638550 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-config-data\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: W1004 11:09:50.651655 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fff34a3_8666_465c_b9bd_6bd6e56c17fe.slice/crio-2305362e5890dc4ba51f3cc2ee329a92c4303f5f0d31322f368b7cf873ebf1d8 WatchSource:0}: Error finding container 2305362e5890dc4ba51f3cc2ee329a92c4303f5f0d31322f368b7cf873ebf1d8: Status 404 returned error can't find the container with id 2305362e5890dc4ba51f3cc2ee329a92c4303f5f0d31322f368b7cf873ebf1d8 Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.652772 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.659709 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6zcx\" (UniqueName: \"kubernetes.io/projected/ac470595-559d-4607-9113-450378bcaf13-kube-api-access-d6zcx\") pod \"nova-cell1-conductor-db-sync-gn24j\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.702821 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.754863 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ee35768-f230-4b06-991a-31b79d2d3d78","Type":"ContainerStarted","Data":"509b096c7c452aed41761c9fa918dbeb15b602b5612e1079ba26facaffaadd78"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.757481 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"56db99fe-c181-450e-b9cc-f8810b6658e8","Type":"ContainerStarted","Data":"da7d7d9adaab0130fe20d0fbe71603f9b1a8babebf020cc0aea96704eb53f5a3"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.762219 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4p4g9" event={"ID":"702dad08-3e1d-4118-9ecd-6c1882b8188d","Type":"ContainerStarted","Data":"b74dbf4be71a1fba3808e336d64557bae7afcd821ff80c44aad9c60a32d5769f"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.762259 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4p4g9" event={"ID":"702dad08-3e1d-4118-9ecd-6c1882b8188d","Type":"ContainerStarted","Data":"dcdaa69f2a841e657b7250ac0fe687d3af1d4025ecfe411acb68695eb9ff6668"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.767644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3fff34a3-8666-465c-b9bd-6bd6e56c17fe","Type":"ContainerStarted","Data":"2305362e5890dc4ba51f3cc2ee329a92c4303f5f0d31322f368b7cf873ebf1d8"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.771402 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f","Type":"ContainerStarted","Data":"72aa688aa0526d362d38494bd9017044b31f9980b7efc632931311497de36836"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.777221 4758 generic.go:334] "Generic (PLEG): container finished" podID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerID="cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349" exitCode=0 Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.777262 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" event={"ID":"c710639c-0dc3-43ba-b084-fee7da532dc8","Type":"ContainerDied","Data":"cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.777286 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" event={"ID":"c710639c-0dc3-43ba-b084-fee7da532dc8","Type":"ContainerStarted","Data":"34eb4f705bec27171cb22c31da010c9182772b85845a008a9acdf12f7b76d586"} Oct 04 11:09:50 crc kubenswrapper[4758]: I1004 11:09:50.780362 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-4p4g9" podStartSLOduration=2.780343908 podStartE2EDuration="2.780343908s" podCreationTimestamp="2025-10-04 11:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:50.77300319 +0000 UTC m=+1168.065654069" watchObservedRunningTime="2025-10-04 11:09:50.780343908 +0000 UTC m=+1168.072994797" Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.280363 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gn24j"] Oct 04 11:09:51 crc kubenswrapper[4758]: W1004 11:09:51.310531 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac470595_559d_4607_9113_450378bcaf13.slice/crio-484d7a1f45a8294be065c4ce01fe969d7a3cea0a6bcd139a9209d5c6ccb8bc6d WatchSource:0}: Error finding container 484d7a1f45a8294be065c4ce01fe969d7a3cea0a6bcd139a9209d5c6ccb8bc6d: Status 404 returned error can't find the container with id 484d7a1f45a8294be065c4ce01fe969d7a3cea0a6bcd139a9209d5c6ccb8bc6d Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.792182 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" event={"ID":"c710639c-0dc3-43ba-b084-fee7da532dc8","Type":"ContainerStarted","Data":"8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83"} Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.792835 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.795547 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gn24j" event={"ID":"ac470595-559d-4607-9113-450378bcaf13","Type":"ContainerStarted","Data":"08bc7f34c9986fcd62327c5e0164d0c73c636de627d599923e049bf336c847f5"} Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.795566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gn24j" event={"ID":"ac470595-559d-4607-9113-450378bcaf13","Type":"ContainerStarted","Data":"484d7a1f45a8294be065c4ce01fe969d7a3cea0a6bcd139a9209d5c6ccb8bc6d"} Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.822538 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" podStartSLOduration=2.822514055 podStartE2EDuration="2.822514055s" podCreationTimestamp="2025-10-04 11:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:51.809782241 +0000 UTC m=+1169.102433130" watchObservedRunningTime="2025-10-04 11:09:51.822514055 +0000 UTC m=+1169.115164944" Oct 04 11:09:51 crc kubenswrapper[4758]: I1004 11:09:51.831982 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gn24j" podStartSLOduration=1.83196505 podStartE2EDuration="1.83196505s" podCreationTimestamp="2025-10-04 11:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:09:51.827946412 +0000 UTC m=+1169.120597301" watchObservedRunningTime="2025-10-04 11:09:51.83196505 +0000 UTC m=+1169.124615939" Oct 04 11:09:52 crc kubenswrapper[4758]: I1004 11:09:52.916598 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:09:52 crc kubenswrapper[4758]: I1004 11:09:52.934886 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.142739 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.143479 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f3b5b5a4-cf2e-489c-acd3-61811589fa03" containerName="kube-state-metrics" containerID="cri-o://c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8" gracePeriod=30 Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.594187 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.777662 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6vbp\" (UniqueName: \"kubernetes.io/projected/f3b5b5a4-cf2e-489c-acd3-61811589fa03-kube-api-access-f6vbp\") pod \"f3b5b5a4-cf2e-489c-acd3-61811589fa03\" (UID: \"f3b5b5a4-cf2e-489c-acd3-61811589fa03\") " Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.781578 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3b5b5a4-cf2e-489c-acd3-61811589fa03-kube-api-access-f6vbp" (OuterVolumeSpecName: "kube-api-access-f6vbp") pod "f3b5b5a4-cf2e-489c-acd3-61811589fa03" (UID: "f3b5b5a4-cf2e-489c-acd3-61811589fa03"). InnerVolumeSpecName "kube-api-access-f6vbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.828697 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"56db99fe-c181-450e-b9cc-f8810b6658e8","Type":"ContainerStarted","Data":"266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.830999 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3fff34a3-8666-465c-b9bd-6bd6e56c17fe","Type":"ContainerStarted","Data":"7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.831053 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3fff34a3-8666-465c-b9bd-6bd6e56c17fe" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5" gracePeriod=30 Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.832853 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f","Type":"ContainerStarted","Data":"418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.832878 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f","Type":"ContainerStarted","Data":"0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.833147 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-log" containerID="cri-o://0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5" gracePeriod=30 Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.833347 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-metadata" containerID="cri-o://418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9" gracePeriod=30 Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.840048 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ee35768-f230-4b06-991a-31b79d2d3d78","Type":"ContainerStarted","Data":"4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.840148 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ee35768-f230-4b06-991a-31b79d2d3d78","Type":"ContainerStarted","Data":"a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.841606 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3b5b5a4-cf2e-489c-acd3-61811589fa03" containerID="c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8" exitCode=2 Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.841644 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3b5b5a4-cf2e-489c-acd3-61811589fa03","Type":"ContainerDied","Data":"c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.841663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f3b5b5a4-cf2e-489c-acd3-61811589fa03","Type":"ContainerDied","Data":"6b3957ceff5117d07881b1bf016c8348b10faa0a76b3759074dc947d88b924a0"} Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.841681 4758 scope.go:117] "RemoveContainer" containerID="c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.841775 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.879331 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.57230176 podStartE2EDuration="6.879310279s" podCreationTimestamp="2025-10-04 11:09:49 +0000 UTC" firstStartedPulling="2025-10-04 11:09:50.654480601 +0000 UTC m=+1167.947131490" lastFinishedPulling="2025-10-04 11:09:54.96148912 +0000 UTC m=+1172.254140009" observedRunningTime="2025-10-04 11:09:55.875437054 +0000 UTC m=+1173.168087943" watchObservedRunningTime="2025-10-04 11:09:55.879310279 +0000 UTC m=+1173.171961158" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.879530 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6vbp\" (UniqueName: \"kubernetes.io/projected/f3b5b5a4-cf2e-489c-acd3-61811589fa03-kube-api-access-f6vbp\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.879952 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.285350952 podStartE2EDuration="6.879947426s" podCreationTimestamp="2025-10-04 11:09:49 +0000 UTC" firstStartedPulling="2025-10-04 11:09:50.360038001 +0000 UTC m=+1167.652688890" lastFinishedPulling="2025-10-04 11:09:54.954634455 +0000 UTC m=+1172.247285364" observedRunningTime="2025-10-04 11:09:55.853356008 +0000 UTC m=+1173.146006897" watchObservedRunningTime="2025-10-04 11:09:55.879947426 +0000 UTC m=+1173.172598315" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.888306 4758 scope.go:117] "RemoveContainer" containerID="c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8" Oct 04 11:09:55 crc kubenswrapper[4758]: E1004 11:09:55.895279 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8\": container with ID starting with c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8 not found: ID does not exist" containerID="c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.895325 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8"} err="failed to get container status \"c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8\": rpc error: code = NotFound desc = could not find container \"c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8\": container with ID starting with c2f96e4ec26a77c9f983ac0f5dcd621e519af01439b65dd129f677970d39bcd8 not found: ID does not exist" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.922172 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.065912287 podStartE2EDuration="7.922143725s" podCreationTimestamp="2025-10-04 11:09:48 +0000 UTC" firstStartedPulling="2025-10-04 11:09:50.105791707 +0000 UTC m=+1167.398442586" lastFinishedPulling="2025-10-04 11:09:54.962023125 +0000 UTC m=+1172.254674024" observedRunningTime="2025-10-04 11:09:55.901860047 +0000 UTC m=+1173.194510936" watchObservedRunningTime="2025-10-04 11:09:55.922143725 +0000 UTC m=+1173.214794614" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.935638 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.500205023 podStartE2EDuration="6.935619719s" podCreationTimestamp="2025-10-04 11:09:49 +0000 UTC" firstStartedPulling="2025-10-04 11:09:50.537820991 +0000 UTC m=+1167.830471870" lastFinishedPulling="2025-10-04 11:09:54.973235667 +0000 UTC m=+1172.265886566" observedRunningTime="2025-10-04 11:09:55.929164015 +0000 UTC m=+1173.221814904" watchObservedRunningTime="2025-10-04 11:09:55.935619719 +0000 UTC m=+1173.228270608" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.955177 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.965998 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.994645 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:09:55 crc kubenswrapper[4758]: E1004 11:09:55.995057 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3b5b5a4-cf2e-489c-acd3-61811589fa03" containerName="kube-state-metrics" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.995074 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3b5b5a4-cf2e-489c-acd3-61811589fa03" containerName="kube-state-metrics" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.995271 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3b5b5a4-cf2e-489c-acd3-61811589fa03" containerName="kube-state-metrics" Oct 04 11:09:55 crc kubenswrapper[4758]: I1004 11:09:55.995864 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.008723 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.008908 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.010720 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.084297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjqzm\" (UniqueName: \"kubernetes.io/projected/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-api-access-kjqzm\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.084349 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.084391 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.084433 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.185584 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.185905 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.186384 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.186499 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjqzm\" (UniqueName: \"kubernetes.io/projected/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-api-access-kjqzm\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.190987 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.191907 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.195612 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.204869 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjqzm\" (UniqueName: \"kubernetes.io/projected/bb9965b2-702d-4b0e-b8ef-2590608e2ebe-kube-api-access-kjqzm\") pod \"kube-state-metrics-0\" (UID: \"bb9965b2-702d-4b0e-b8ef-2590608e2ebe\") " pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.367220 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.904509 4758 generic.go:334] "Generic (PLEG): container finished" podID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerID="0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5" exitCode=143 Oct 04 11:09:56 crc kubenswrapper[4758]: I1004 11:09:56.904835 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f","Type":"ContainerDied","Data":"0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5"} Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.079200 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.340143 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3b5b5a4-cf2e-489c-acd3-61811589fa03" path="/var/lib/kubelet/pods/f3b5b5a4-cf2e-489c-acd3-61811589fa03/volumes" Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.401149 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.401652 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="sg-core" containerID="cri-o://b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56" gracePeriod=30 Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.401478 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-central-agent" containerID="cri-o://84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483" gracePeriod=30 Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.401685 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="proxy-httpd" containerID="cri-o://eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099" gracePeriod=30 Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.401707 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-notification-agent" containerID="cri-o://43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6" gracePeriod=30 Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.927553 4758 generic.go:334] "Generic (PLEG): container finished" podID="3520d415-f35c-44f0-a001-638478475b44" containerID="eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099" exitCode=0 Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.927588 4758 generic.go:334] "Generic (PLEG): container finished" podID="3520d415-f35c-44f0-a001-638478475b44" containerID="b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56" exitCode=2 Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.927645 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerDied","Data":"eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099"} Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.927675 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerDied","Data":"b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56"} Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.929595 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bb9965b2-702d-4b0e-b8ef-2590608e2ebe","Type":"ContainerStarted","Data":"9529ff142b7b49e52505cdaa411001cb79b713ef77d8d8c8b37bef78c4f74204"} Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.929628 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"bb9965b2-702d-4b0e-b8ef-2590608e2ebe","Type":"ContainerStarted","Data":"c6a2cbeffc1c7588b159198d3646ee8f14f66e84e307a1211009c2f6e140a024"} Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.930013 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 11:09:57 crc kubenswrapper[4758]: I1004 11:09:57.946232 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.594787653 podStartE2EDuration="2.946215931s" podCreationTimestamp="2025-10-04 11:09:55 +0000 UTC" firstStartedPulling="2025-10-04 11:09:57.08316048 +0000 UTC m=+1174.375811369" lastFinishedPulling="2025-10-04 11:09:57.434588758 +0000 UTC m=+1174.727239647" observedRunningTime="2025-10-04 11:09:57.945224454 +0000 UTC m=+1175.237875343" watchObservedRunningTime="2025-10-04 11:09:57.946215931 +0000 UTC m=+1175.238866820" Oct 04 11:09:58 crc kubenswrapper[4758]: I1004 11:09:58.573537 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-549f9cdcb8-66q22" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 04 11:09:58 crc kubenswrapper[4758]: I1004 11:09:58.573959 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:09:58 crc kubenswrapper[4758]: I1004 11:09:58.983475 4758 generic.go:334] "Generic (PLEG): container finished" podID="3520d415-f35c-44f0-a001-638478475b44" containerID="84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483" exitCode=0 Oct 04 11:09:58 crc kubenswrapper[4758]: I1004 11:09:58.983567 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerDied","Data":"84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483"} Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.407891 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.407968 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.578598 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.578986 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.592642 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.626161 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.683211 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kt56w"] Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.683443 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerName="dnsmasq-dns" containerID="cri-o://eb11b7a9820a6929a2648090834f93daf33a7376779b897f063a5b27b3c74b93" gracePeriod=10 Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.774552 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.774599 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:09:59 crc kubenswrapper[4758]: I1004 11:09:59.852317 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.002867 4758 generic.go:334] "Generic (PLEG): container finished" podID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerID="eb11b7a9820a6929a2648090834f93daf33a7376779b897f063a5b27b3c74b93" exitCode=0 Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.002934 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" event={"ID":"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7","Type":"ContainerDied","Data":"eb11b7a9820a6929a2648090834f93daf33a7376779b897f063a5b27b3c74b93"} Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.015435 4758 generic.go:334] "Generic (PLEG): container finished" podID="702dad08-3e1d-4118-9ecd-6c1882b8188d" containerID="b74dbf4be71a1fba3808e336d64557bae7afcd821ff80c44aad9c60a32d5769f" exitCode=0 Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.015785 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4p4g9" event={"ID":"702dad08-3e1d-4118-9ecd-6c1882b8188d","Type":"ContainerDied","Data":"b74dbf4be71a1fba3808e336d64557bae7afcd821ff80c44aad9c60a32d5769f"} Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.059127 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.170890 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.289424 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz4rc\" (UniqueName: \"kubernetes.io/projected/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-kube-api-access-qz4rc\") pod \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.289481 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-svc\") pod \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.289523 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-nb\") pod \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.289555 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-sb\") pod \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.289606 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-config\") pod \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.289730 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-swift-storage-0\") pod \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\" (UID: \"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7\") " Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.298958 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-kube-api-access-qz4rc" (OuterVolumeSpecName: "kube-api-access-qz4rc") pod "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" (UID: "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7"). InnerVolumeSpecName "kube-api-access-qz4rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.343825 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" (UID: "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.363902 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" (UID: "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.369659 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" (UID: "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.391535 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.391569 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz4rc\" (UniqueName: \"kubernetes.io/projected/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-kube-api-access-qz4rc\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.391579 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.391589 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.412582 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" (UID: "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.412883 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-config" (OuterVolumeSpecName: "config") pod "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" (UID: "8b34463b-2edd-4bdf-be2d-68b5b9fa03a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.493773 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.494545 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.858658 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:00 crc kubenswrapper[4758]: I1004 11:10:00.858847 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.024909 4758 generic.go:334] "Generic (PLEG): container finished" podID="ac470595-559d-4607-9113-450378bcaf13" containerID="08bc7f34c9986fcd62327c5e0164d0c73c636de627d599923e049bf336c847f5" exitCode=0 Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.025014 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gn24j" event={"ID":"ac470595-559d-4607-9113-450378bcaf13","Type":"ContainerDied","Data":"08bc7f34c9986fcd62327c5e0164d0c73c636de627d599923e049bf336c847f5"} Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.026896 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" event={"ID":"8b34463b-2edd-4bdf-be2d-68b5b9fa03a7","Type":"ContainerDied","Data":"b8a3bbe4a74470d6a1ec5495324ee0d2e9f0154bcfe295b824746a17c1a4cb67"} Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.026947 4758 scope.go:117] "RemoveContainer" containerID="eb11b7a9820a6929a2648090834f93daf33a7376779b897f063a5b27b3c74b93" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.027656 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-kt56w" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.075701 4758 scope.go:117] "RemoveContainer" containerID="63753f506ade0576037bc099f5bf3b701c1f6e79c6247ab651a4250d1f075b36" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.076870 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kt56w"] Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.083979 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-kt56w"] Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.249779 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.249828 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.340641 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" path="/var/lib/kubelet/pods/8b34463b-2edd-4bdf-be2d-68b5b9fa03a7/volumes" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.428457 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.612835 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-scripts\") pod \"702dad08-3e1d-4118-9ecd-6c1882b8188d\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.612965 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-config-data\") pod \"702dad08-3e1d-4118-9ecd-6c1882b8188d\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.613029 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zk66\" (UniqueName: \"kubernetes.io/projected/702dad08-3e1d-4118-9ecd-6c1882b8188d-kube-api-access-2zk66\") pod \"702dad08-3e1d-4118-9ecd-6c1882b8188d\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.613056 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-combined-ca-bundle\") pod \"702dad08-3e1d-4118-9ecd-6c1882b8188d\" (UID: \"702dad08-3e1d-4118-9ecd-6c1882b8188d\") " Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.633821 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-scripts" (OuterVolumeSpecName: "scripts") pod "702dad08-3e1d-4118-9ecd-6c1882b8188d" (UID: "702dad08-3e1d-4118-9ecd-6c1882b8188d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.634210 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702dad08-3e1d-4118-9ecd-6c1882b8188d-kube-api-access-2zk66" (OuterVolumeSpecName: "kube-api-access-2zk66") pod "702dad08-3e1d-4118-9ecd-6c1882b8188d" (UID: "702dad08-3e1d-4118-9ecd-6c1882b8188d"). InnerVolumeSpecName "kube-api-access-2zk66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.638259 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "702dad08-3e1d-4118-9ecd-6c1882b8188d" (UID: "702dad08-3e1d-4118-9ecd-6c1882b8188d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.650248 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-config-data" (OuterVolumeSpecName: "config-data") pod "702dad08-3e1d-4118-9ecd-6c1882b8188d" (UID: "702dad08-3e1d-4118-9ecd-6c1882b8188d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.716437 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.716485 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zk66\" (UniqueName: \"kubernetes.io/projected/702dad08-3e1d-4118-9ecd-6c1882b8188d-kube-api-access-2zk66\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.716501 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:01 crc kubenswrapper[4758]: I1004 11:10:01.716512 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702dad08-3e1d-4118-9ecd-6c1882b8188d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.040910 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4p4g9" event={"ID":"702dad08-3e1d-4118-9ecd-6c1882b8188d","Type":"ContainerDied","Data":"dcdaa69f2a841e657b7250ac0fe687d3af1d4025ecfe411acb68695eb9ff6668"} Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.041340 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcdaa69f2a841e657b7250ac0fe687d3af1d4025ecfe411acb68695eb9ff6668" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.040963 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4p4g9" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.240245 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.240521 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-log" containerID="cri-o://a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8" gracePeriod=30 Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.241000 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-api" containerID="cri-o://4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431" gracePeriod=30 Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.253497 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.253931 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="56db99fe-c181-450e-b9cc-f8810b6658e8" containerName="nova-scheduler-scheduler" containerID="cri-o://266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9" gracePeriod=30 Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.501843 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:10:02 crc kubenswrapper[4758]: E1004 11:10:02.502315 4758 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ee35768_f230_4b06_991a_31b79d2d3d78.slice/crio-conmon-a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8.scope\": RecentStats: unable to find data in memory cache]" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.631776 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-combined-ca-bundle\") pod \"ac470595-559d-4607-9113-450378bcaf13\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.631822 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6zcx\" (UniqueName: \"kubernetes.io/projected/ac470595-559d-4607-9113-450378bcaf13-kube-api-access-d6zcx\") pod \"ac470595-559d-4607-9113-450378bcaf13\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.632044 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-scripts\") pod \"ac470595-559d-4607-9113-450378bcaf13\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.632085 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-config-data\") pod \"ac470595-559d-4607-9113-450378bcaf13\" (UID: \"ac470595-559d-4607-9113-450378bcaf13\") " Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.638190 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-scripts" (OuterVolumeSpecName: "scripts") pod "ac470595-559d-4607-9113-450378bcaf13" (UID: "ac470595-559d-4607-9113-450378bcaf13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.653299 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac470595-559d-4607-9113-450378bcaf13-kube-api-access-d6zcx" (OuterVolumeSpecName: "kube-api-access-d6zcx") pod "ac470595-559d-4607-9113-450378bcaf13" (UID: "ac470595-559d-4607-9113-450378bcaf13"). InnerVolumeSpecName "kube-api-access-d6zcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.669557 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-config-data" (OuterVolumeSpecName: "config-data") pod "ac470595-559d-4607-9113-450378bcaf13" (UID: "ac470595-559d-4607-9113-450378bcaf13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.671252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac470595-559d-4607-9113-450378bcaf13" (UID: "ac470595-559d-4607-9113-450378bcaf13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.733933 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.733968 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6zcx\" (UniqueName: \"kubernetes.io/projected/ac470595-559d-4607-9113-450378bcaf13-kube-api-access-d6zcx\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.733981 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:02 crc kubenswrapper[4758]: I1004 11:10:02.733989 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac470595-559d-4607-9113-450378bcaf13-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.064844 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gn24j" event={"ID":"ac470595-559d-4607-9113-450378bcaf13","Type":"ContainerDied","Data":"484d7a1f45a8294be065c4ce01fe969d7a3cea0a6bcd139a9209d5c6ccb8bc6d"} Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.065151 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="484d7a1f45a8294be065c4ce01fe969d7a3cea0a6bcd139a9209d5c6ccb8bc6d" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.065206 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gn24j" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.091596 4758 generic.go:334] "Generic (PLEG): container finished" podID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerID="a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8" exitCode=143 Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.091681 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ee35768-f230-4b06-991a-31b79d2d3d78","Type":"ContainerDied","Data":"a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8"} Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.106765 4758 generic.go:334] "Generic (PLEG): container finished" podID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerID="e10f41f31bbb4f5fd1672270aaecda9dd129769e4c72c63717d0592b51d67711" exitCode=137 Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.106800 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549f9cdcb8-66q22" event={"ID":"55a162df-bf26-4c9f-9808-506db4c3bd01","Type":"ContainerDied","Data":"e10f41f31bbb4f5fd1672270aaecda9dd129769e4c72c63717d0592b51d67711"} Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.140159 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 11:10:03 crc kubenswrapper[4758]: E1004 11:10:03.140483 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac470595-559d-4607-9113-450378bcaf13" containerName="nova-cell1-conductor-db-sync" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.140500 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac470595-559d-4607-9113-450378bcaf13" containerName="nova-cell1-conductor-db-sync" Oct 04 11:10:03 crc kubenswrapper[4758]: E1004 11:10:03.140520 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerName="dnsmasq-dns" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.140527 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerName="dnsmasq-dns" Oct 04 11:10:03 crc kubenswrapper[4758]: E1004 11:10:03.140539 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerName="init" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.140545 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerName="init" Oct 04 11:10:03 crc kubenswrapper[4758]: E1004 11:10:03.140556 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702dad08-3e1d-4118-9ecd-6c1882b8188d" containerName="nova-manage" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.140563 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="702dad08-3e1d-4118-9ecd-6c1882b8188d" containerName="nova-manage" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.177622 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="702dad08-3e1d-4118-9ecd-6c1882b8188d" containerName="nova-manage" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.177675 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b34463b-2edd-4bdf-be2d-68b5b9fa03a7" containerName="dnsmasq-dns" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.177693 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac470595-559d-4607-9113-450378bcaf13" containerName="nova-cell1-conductor-db-sync" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.185053 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.187974 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.207093 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.259385 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.355851 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nst7\" (UniqueName: \"kubernetes.io/projected/1a824dc0-f283-4e65-9b90-82b0d55e478d-kube-api-access-7nst7\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.355924 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a824dc0-f283-4e65-9b90-82b0d55e478d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.355955 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a824dc0-f283-4e65-9b90-82b0d55e478d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457619 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-config-data\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457745 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-secret-key\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457783 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-combined-ca-bundle\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457805 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-tls-certs\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457823 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55a162df-bf26-4c9f-9808-506db4c3bd01-logs\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457872 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-scripts\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.457919 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phpw5\" (UniqueName: \"kubernetes.io/projected/55a162df-bf26-4c9f-9808-506db4c3bd01-kube-api-access-phpw5\") pod \"55a162df-bf26-4c9f-9808-506db4c3bd01\" (UID: \"55a162df-bf26-4c9f-9808-506db4c3bd01\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.458254 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nst7\" (UniqueName: \"kubernetes.io/projected/1a824dc0-f283-4e65-9b90-82b0d55e478d-kube-api-access-7nst7\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.458299 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a824dc0-f283-4e65-9b90-82b0d55e478d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.458330 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a824dc0-f283-4e65-9b90-82b0d55e478d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.459435 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a162df-bf26-4c9f-9808-506db4c3bd01-logs" (OuterVolumeSpecName: "logs") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.464795 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a824dc0-f283-4e65-9b90-82b0d55e478d-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.476620 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a162df-bf26-4c9f-9808-506db4c3bd01-kube-api-access-phpw5" (OuterVolumeSpecName: "kube-api-access-phpw5") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "kube-api-access-phpw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.477626 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.478863 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a824dc0-f283-4e65-9b90-82b0d55e478d-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.500770 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nst7\" (UniqueName: \"kubernetes.io/projected/1a824dc0-f283-4e65-9b90-82b0d55e478d-kube-api-access-7nst7\") pod \"nova-cell1-conductor-0\" (UID: \"1a824dc0-f283-4e65-9b90-82b0d55e478d\") " pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.514244 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-scripts" (OuterVolumeSpecName: "scripts") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.520667 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-config-data" (OuterVolumeSpecName: "config-data") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.527844 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.547892 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "55a162df-bf26-4c9f-9808-506db4c3bd01" (UID: "55a162df-bf26-4c9f-9808-506db4c3bd01"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560527 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560650 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560935 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phpw5\" (UniqueName: \"kubernetes.io/projected/55a162df-bf26-4c9f-9808-506db4c3bd01-kube-api-access-phpw5\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560956 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55a162df-bf26-4c9f-9808-506db4c3bd01-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560966 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560976 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560985 4758 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55a162df-bf26-4c9f-9808-506db4c3bd01-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.560993 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55a162df-bf26-4c9f-9808-506db4c3bd01-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.612839 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-config-data\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764279 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mwzk\" (UniqueName: \"kubernetes.io/projected/3520d415-f35c-44f0-a001-638478475b44-kube-api-access-2mwzk\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764325 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-scripts\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764401 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-combined-ca-bundle\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764432 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-log-httpd\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764476 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-sg-core-conf-yaml\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764519 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-run-httpd\") pod \"3520d415-f35c-44f0-a001-638478475b44\" (UID: \"3520d415-f35c-44f0-a001-638478475b44\") " Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.764862 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.765125 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.769359 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-scripts" (OuterVolumeSpecName: "scripts") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.770819 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3520d415-f35c-44f0-a001-638478475b44-kube-api-access-2mwzk" (OuterVolumeSpecName: "kube-api-access-2mwzk") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "kube-api-access-2mwzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.827891 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.868940 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mwzk\" (UniqueName: \"kubernetes.io/projected/3520d415-f35c-44f0-a001-638478475b44-kube-api-access-2mwzk\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.869765 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.869777 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.869786 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.869794 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3520d415-f35c-44f0-a001-638478475b44-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.920865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-config-data" (OuterVolumeSpecName: "config-data") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.930274 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3520d415-f35c-44f0-a001-638478475b44" (UID: "3520d415-f35c-44f0-a001-638478475b44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.972744 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:03 crc kubenswrapper[4758]: I1004 11:10:03.972774 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3520d415-f35c-44f0-a001-638478475b44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.044935 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.116625 4758 generic.go:334] "Generic (PLEG): container finished" podID="56db99fe-c181-450e-b9cc-f8810b6658e8" containerID="266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9" exitCode=0 Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.116732 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"56db99fe-c181-450e-b9cc-f8810b6658e8","Type":"ContainerDied","Data":"266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9"} Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.116761 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"56db99fe-c181-450e-b9cc-f8810b6658e8","Type":"ContainerDied","Data":"da7d7d9adaab0130fe20d0fbe71603f9b1a8babebf020cc0aea96704eb53f5a3"} Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.116779 4758 scope.go:117] "RemoveContainer" containerID="266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.118167 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.119751 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549f9cdcb8-66q22" event={"ID":"55a162df-bf26-4c9f-9808-506db4c3bd01","Type":"ContainerDied","Data":"48a04c4673475bdee1e0e3b82de7092eb94182c76e58803f5cfcb28ab1c9cac9"} Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.119754 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549f9cdcb8-66q22" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.122554 4758 generic.go:334] "Generic (PLEG): container finished" podID="3520d415-f35c-44f0-a001-638478475b44" containerID="43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6" exitCode=0 Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.122684 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerDied","Data":"43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6"} Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.122871 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3520d415-f35c-44f0-a001-638478475b44","Type":"ContainerDied","Data":"9666266804d214cc078e4f2cba765d4789c776b0b390311394a8c653a65f0cbb"} Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.122834 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.142315 4758 scope.go:117] "RemoveContainer" containerID="266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.148882 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9\": container with ID starting with 266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9 not found: ID does not exist" containerID="266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.149073 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9"} err="failed to get container status \"266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9\": rpc error: code = NotFound desc = could not find container \"266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9\": container with ID starting with 266d0f78f056c4f354b52d203fd93145686c75b9bf75511d534bf338939906a9 not found: ID does not exist" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.149190 4758 scope.go:117] "RemoveContainer" containerID="b7c2b7c398b965f1af57612a5dd546394a6cac04c79b30a09d3ac182e4e98488" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.172409 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.178575 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-config-data\") pod \"56db99fe-c181-450e-b9cc-f8810b6658e8\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.178999 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2wss\" (UniqueName: \"kubernetes.io/projected/56db99fe-c181-450e-b9cc-f8810b6658e8-kube-api-access-v2wss\") pod \"56db99fe-c181-450e-b9cc-f8810b6658e8\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.179614 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-combined-ca-bundle\") pod \"56db99fe-c181-450e-b9cc-f8810b6658e8\" (UID: \"56db99fe-c181-450e-b9cc-f8810b6658e8\") " Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.192860 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.202078 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56db99fe-c181-450e-b9cc-f8810b6658e8-kube-api-access-v2wss" (OuterVolumeSpecName: "kube-api-access-v2wss") pod "56db99fe-c181-450e-b9cc-f8810b6658e8" (UID: "56db99fe-c181-450e-b9cc-f8810b6658e8"). InnerVolumeSpecName "kube-api-access-v2wss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.203380 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.213624 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214045 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon-log" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214064 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon-log" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214080 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214087 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214111 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56db99fe-c181-450e-b9cc-f8810b6658e8" containerName="nova-scheduler-scheduler" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214146 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="56db99fe-c181-450e-b9cc-f8810b6658e8" containerName="nova-scheduler-scheduler" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214177 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="proxy-httpd" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214183 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="proxy-httpd" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214193 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-central-agent" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214199 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-central-agent" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214206 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-notification-agent" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214234 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-notification-agent" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.214257 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="sg-core" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214265 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="sg-core" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214598 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-central-agent" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214623 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="56db99fe-c181-450e-b9cc-f8810b6658e8" containerName="nova-scheduler-scheduler" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214631 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="sg-core" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214642 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon-log" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214681 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="proxy-httpd" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214702 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" containerName="horizon" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.214712 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3520d415-f35c-44f0-a001-638478475b44" containerName="ceilometer-notification-agent" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.216312 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.224486 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.224684 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.224794 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.227227 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-549f9cdcb8-66q22"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.237532 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-549f9cdcb8-66q22"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.251627 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.252332 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-config-data" (OuterVolumeSpecName: "config-data") pod "56db99fe-c181-450e-b9cc-f8810b6658e8" (UID: "56db99fe-c181-450e-b9cc-f8810b6658e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.256141 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56db99fe-c181-450e-b9cc-f8810b6658e8" (UID: "56db99fe-c181-450e-b9cc-f8810b6658e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.282017 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.282046 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2wss\" (UniqueName: \"kubernetes.io/projected/56db99fe-c181-450e-b9cc-f8810b6658e8-kube-api-access-v2wss\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.282055 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56db99fe-c181-450e-b9cc-f8810b6658e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.382991 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.383432 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.383481 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.383523 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-config-data\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.383649 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-run-httpd\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.383823 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-scripts\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.383900 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-log-httpd\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.384010 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqmgm\" (UniqueName: \"kubernetes.io/projected/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-kube-api-access-tqmgm\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.404695 4758 scope.go:117] "RemoveContainer" containerID="e10f41f31bbb4f5fd1672270aaecda9dd129769e4c72c63717d0592b51d67711" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.449966 4758 scope.go:117] "RemoveContainer" containerID="eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.453505 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.466798 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.479518 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.480660 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.480788 4758 scope.go:117] "RemoveContainer" containerID="b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.482598 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485275 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485427 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485523 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-config-data\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485567 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-run-httpd\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485605 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-scripts\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485629 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-log-httpd\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.485648 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqmgm\" (UniqueName: \"kubernetes.io/projected/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-kube-api-access-tqmgm\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.487902 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-log-httpd\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.493871 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-run-httpd\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.494876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.499549 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-scripts\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.500232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.502147 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.508679 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-config-data\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.519191 4758 scope.go:117] "RemoveContainer" containerID="43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.531510 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.562520 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqmgm\" (UniqueName: \"kubernetes.io/projected/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-kube-api-access-tqmgm\") pod \"ceilometer-0\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.563894 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.587735 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-config-data\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.587789 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb4gb\" (UniqueName: \"kubernetes.io/projected/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-kube-api-access-kb4gb\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.587885 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.603985 4758 scope.go:117] "RemoveContainer" containerID="84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.630726 4758 scope.go:117] "RemoveContainer" containerID="eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.633487 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099\": container with ID starting with eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099 not found: ID does not exist" containerID="eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.633530 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099"} err="failed to get container status \"eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099\": rpc error: code = NotFound desc = could not find container \"eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099\": container with ID starting with eed0c951c17e0e7011e3a6a3eb1c04a9e7326c490950fa368fa05beeea6d5099 not found: ID does not exist" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.633557 4758 scope.go:117] "RemoveContainer" containerID="b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.633964 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56\": container with ID starting with b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56 not found: ID does not exist" containerID="b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.633988 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56"} err="failed to get container status \"b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56\": rpc error: code = NotFound desc = could not find container \"b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56\": container with ID starting with b42c0dd80d79b43631eb66c2fcef1b4dbc766efc2d53b730185ed8a24dbc4a56 not found: ID does not exist" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.634003 4758 scope.go:117] "RemoveContainer" containerID="43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.634819 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6\": container with ID starting with 43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6 not found: ID does not exist" containerID="43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.634843 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6"} err="failed to get container status \"43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6\": rpc error: code = NotFound desc = could not find container \"43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6\": container with ID starting with 43cde5474e252e45ab7a9666a1eb31c125321698427144508f818a21ee33aab6 not found: ID does not exist" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.634857 4758 scope.go:117] "RemoveContainer" containerID="84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483" Oct 04 11:10:04 crc kubenswrapper[4758]: E1004 11:10:04.635234 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483\": container with ID starting with 84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483 not found: ID does not exist" containerID="84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.635255 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483"} err="failed to get container status \"84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483\": rpc error: code = NotFound desc = could not find container \"84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483\": container with ID starting with 84735d4f94a8a60b3af1dbbf5b89d2c6776975641337019ffe690255aa9d0483 not found: ID does not exist" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.690171 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-config-data\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.690226 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb4gb\" (UniqueName: \"kubernetes.io/projected/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-kube-api-access-kb4gb\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.690315 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.693784 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-config-data\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.696393 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.712584 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb4gb\" (UniqueName: \"kubernetes.io/projected/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-kube-api-access-kb4gb\") pod \"nova-scheduler-0\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:04 crc kubenswrapper[4758]: I1004 11:10:04.819534 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.041114 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.135539 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerStarted","Data":"58282a3b14d402b604ec6ed41d52729cb15dfec963b08909166fbc20b200614b"} Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.146884 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1a824dc0-f283-4e65-9b90-82b0d55e478d","Type":"ContainerStarted","Data":"dfa14848166773592717f3765deffb15ec7d316c4eee5e149e197a69678dbb04"} Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.146923 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1a824dc0-f283-4e65-9b90-82b0d55e478d","Type":"ContainerStarted","Data":"b8244a81f329321058defab0eac6f7167534ea1c7ffdb2a49c510f799f20a902"} Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.147238 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.222594 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.222575236 podStartE2EDuration="2.222575236s" podCreationTimestamp="2025-10-04 11:10:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:05.162727951 +0000 UTC m=+1182.455378850" watchObservedRunningTime="2025-10-04 11:10:05.222575236 +0000 UTC m=+1182.515226125" Oct 04 11:10:05 crc kubenswrapper[4758]: W1004 11:10:05.228845 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a73c1d4_ecc4_452c_85a1_c0f8d0fc8460.slice/crio-bdb3eefb404add15bb877407215b510afdd5b0ea58e1397fd3349bd9f0beab56 WatchSource:0}: Error finding container bdb3eefb404add15bb877407215b510afdd5b0ea58e1397fd3349bd9f0beab56: Status 404 returned error can't find the container with id bdb3eefb404add15bb877407215b510afdd5b0ea58e1397fd3349bd9f0beab56 Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.234700 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.337045 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3520d415-f35c-44f0-a001-638478475b44" path="/var/lib/kubelet/pods/3520d415-f35c-44f0-a001-638478475b44/volumes" Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.338020 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a162df-bf26-4c9f-9808-506db4c3bd01" path="/var/lib/kubelet/pods/55a162df-bf26-4c9f-9808-506db4c3bd01/volumes" Oct 04 11:10:05 crc kubenswrapper[4758]: I1004 11:10:05.339503 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56db99fe-c181-450e-b9cc-f8810b6658e8" path="/var/lib/kubelet/pods/56db99fe-c181-450e-b9cc-f8810b6658e8/volumes" Oct 04 11:10:06 crc kubenswrapper[4758]: I1004 11:10:06.154869 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460","Type":"ContainerStarted","Data":"d4a16210f529fcab2281e66777f877df4dccd8bc8b7a7fbf5d4a5d9d577f5136"} Oct 04 11:10:06 crc kubenswrapper[4758]: I1004 11:10:06.155188 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460","Type":"ContainerStarted","Data":"bdb3eefb404add15bb877407215b510afdd5b0ea58e1397fd3349bd9f0beab56"} Oct 04 11:10:06 crc kubenswrapper[4758]: I1004 11:10:06.158992 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerStarted","Data":"4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c"} Oct 04 11:10:06 crc kubenswrapper[4758]: I1004 11:10:06.171932 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.171914347 podStartE2EDuration="2.171914347s" podCreationTimestamp="2025-10-04 11:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:06.170051946 +0000 UTC m=+1183.462702835" watchObservedRunningTime="2025-10-04 11:10:06.171914347 +0000 UTC m=+1183.464565236" Oct 04 11:10:06 crc kubenswrapper[4758]: I1004 11:10:06.387424 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.085344 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.155781 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-config-data\") pod \"0ee35768-f230-4b06-991a-31b79d2d3d78\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.156935 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-combined-ca-bundle\") pod \"0ee35768-f230-4b06-991a-31b79d2d3d78\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.159856 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ee35768-f230-4b06-991a-31b79d2d3d78-logs\") pod \"0ee35768-f230-4b06-991a-31b79d2d3d78\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.160052 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxdpw\" (UniqueName: \"kubernetes.io/projected/0ee35768-f230-4b06-991a-31b79d2d3d78-kube-api-access-jxdpw\") pod \"0ee35768-f230-4b06-991a-31b79d2d3d78\" (UID: \"0ee35768-f230-4b06-991a-31b79d2d3d78\") " Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.160520 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ee35768-f230-4b06-991a-31b79d2d3d78-logs" (OuterVolumeSpecName: "logs") pod "0ee35768-f230-4b06-991a-31b79d2d3d78" (UID: "0ee35768-f230-4b06-991a-31b79d2d3d78"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.175438 4758 generic.go:334] "Generic (PLEG): container finished" podID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerID="4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431" exitCode=0 Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.175499 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ee35768-f230-4b06-991a-31b79d2d3d78","Type":"ContainerDied","Data":"4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431"} Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.175523 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0ee35768-f230-4b06-991a-31b79d2d3d78","Type":"ContainerDied","Data":"509b096c7c452aed41761c9fa918dbeb15b602b5612e1079ba26facaffaadd78"} Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.175540 4758 scope.go:117] "RemoveContainer" containerID="4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.175654 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.190620 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ee35768-f230-4b06-991a-31b79d2d3d78-kube-api-access-jxdpw" (OuterVolumeSpecName: "kube-api-access-jxdpw") pod "0ee35768-f230-4b06-991a-31b79d2d3d78" (UID: "0ee35768-f230-4b06-991a-31b79d2d3d78"). InnerVolumeSpecName "kube-api-access-jxdpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.207544 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ee35768-f230-4b06-991a-31b79d2d3d78" (UID: "0ee35768-f230-4b06-991a-31b79d2d3d78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.207937 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerStarted","Data":"55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc"} Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.210437 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-config-data" (OuterVolumeSpecName: "config-data") pod "0ee35768-f230-4b06-991a-31b79d2d3d78" (UID: "0ee35768-f230-4b06-991a-31b79d2d3d78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.261289 4758 scope.go:117] "RemoveContainer" containerID="a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.266164 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxdpw\" (UniqueName: \"kubernetes.io/projected/0ee35768-f230-4b06-991a-31b79d2d3d78-kube-api-access-jxdpw\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.266199 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.266208 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ee35768-f230-4b06-991a-31b79d2d3d78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.266217 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ee35768-f230-4b06-991a-31b79d2d3d78-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.295983 4758 scope.go:117] "RemoveContainer" containerID="4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431" Oct 04 11:10:07 crc kubenswrapper[4758]: E1004 11:10:07.297486 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431\": container with ID starting with 4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431 not found: ID does not exist" containerID="4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.297542 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431"} err="failed to get container status \"4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431\": rpc error: code = NotFound desc = could not find container \"4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431\": container with ID starting with 4c93dcfa34cf13d1eead3b9c6c552c8f9b6ea7f39305767741ab3fbf4f777431 not found: ID does not exist" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.297583 4758 scope.go:117] "RemoveContainer" containerID="a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8" Oct 04 11:10:07 crc kubenswrapper[4758]: E1004 11:10:07.297996 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8\": container with ID starting with a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8 not found: ID does not exist" containerID="a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.298031 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8"} err="failed to get container status \"a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8\": rpc error: code = NotFound desc = could not find container \"a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8\": container with ID starting with a9d21cdf792fefc79a1a426184c959abd2dc71f6cd4600c01d1fb7781adb1ea8 not found: ID does not exist" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.497815 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.512071 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.526594 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:07 crc kubenswrapper[4758]: E1004 11:10:07.527208 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-log" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.527274 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-log" Oct 04 11:10:07 crc kubenswrapper[4758]: E1004 11:10:07.527343 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-api" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.527394 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-api" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.527633 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-log" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.527705 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" containerName="nova-api-api" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.529711 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.534408 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.538649 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.573591 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-logs\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.573696 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-config-data\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.573778 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9tq2\" (UniqueName: \"kubernetes.io/projected/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-kube-api-access-z9tq2\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.573900 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.675443 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-logs\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.675504 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-config-data\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.675540 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9tq2\" (UniqueName: \"kubernetes.io/projected/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-kube-api-access-z9tq2\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.675594 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.675911 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-logs\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.679499 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-config-data\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.679968 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.692792 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9tq2\" (UniqueName: \"kubernetes.io/projected/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-kube-api-access-z9tq2\") pod \"nova-api-0\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " pod="openstack/nova-api-0" Oct 04 11:10:07 crc kubenswrapper[4758]: I1004 11:10:07.860854 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:08 crc kubenswrapper[4758]: I1004 11:10:08.230374 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerStarted","Data":"a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581"} Oct 04 11:10:08 crc kubenswrapper[4758]: I1004 11:10:08.393172 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.238996 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a6ab03a-ce43-4cf7-a03b-716538b8b26e","Type":"ContainerStarted","Data":"5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124"} Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.239479 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a6ab03a-ce43-4cf7-a03b-716538b8b26e","Type":"ContainerStarted","Data":"9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319"} Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.239490 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a6ab03a-ce43-4cf7-a03b-716538b8b26e","Type":"ContainerStarted","Data":"364b97701067343c7e792fef4e6b3deb8910954cd5486f2a221384a8e56c0a38"} Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.242352 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerStarted","Data":"a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc"} Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.242532 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.259457 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.259441433 podStartE2EDuration="2.259441433s" podCreationTimestamp="2025-10-04 11:10:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:09.252551437 +0000 UTC m=+1186.545202326" watchObservedRunningTime="2025-10-04 11:10:09.259441433 +0000 UTC m=+1186.552092322" Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.279784 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.02462514 podStartE2EDuration="5.279767041s" podCreationTimestamp="2025-10-04 11:10:04 +0000 UTC" firstStartedPulling="2025-10-04 11:10:05.061356084 +0000 UTC m=+1182.354006973" lastFinishedPulling="2025-10-04 11:10:08.316497985 +0000 UTC m=+1185.609148874" observedRunningTime="2025-10-04 11:10:09.269860014 +0000 UTC m=+1186.562510903" watchObservedRunningTime="2025-10-04 11:10:09.279767041 +0000 UTC m=+1186.572417930" Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.334685 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ee35768-f230-4b06-991a-31b79d2d3d78" path="/var/lib/kubelet/pods/0ee35768-f230-4b06-991a-31b79d2d3d78/volumes" Oct 04 11:10:09 crc kubenswrapper[4758]: I1004 11:10:09.820301 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 11:10:13 crc kubenswrapper[4758]: I1004 11:10:13.590349 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 11:10:14 crc kubenswrapper[4758]: I1004 11:10:14.820303 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 11:10:14 crc kubenswrapper[4758]: I1004 11:10:14.853578 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 11:10:15 crc kubenswrapper[4758]: I1004 11:10:15.356328 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 11:10:17 crc kubenswrapper[4758]: I1004 11:10:17.862047 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:10:17 crc kubenswrapper[4758]: I1004 11:10:17.862509 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:10:18 crc kubenswrapper[4758]: I1004 11:10:18.943281 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:18 crc kubenswrapper[4758]: I1004 11:10:18.943542 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.251234 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.361503 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.422821 4758 generic.go:334] "Generic (PLEG): container finished" podID="3fff34a3-8666-465c-b9bd-6bd6e56c17fe" containerID="7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5" exitCode=137 Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.422869 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.422873 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3fff34a3-8666-465c-b9bd-6bd6e56c17fe","Type":"ContainerDied","Data":"7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5"} Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.422936 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3fff34a3-8666-465c-b9bd-6bd6e56c17fe","Type":"ContainerDied","Data":"2305362e5890dc4ba51f3cc2ee329a92c4303f5f0d31322f368b7cf873ebf1d8"} Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.422955 4758 scope.go:117] "RemoveContainer" containerID="7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.425498 4758 generic.go:334] "Generic (PLEG): container finished" podID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerID="418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9" exitCode=137 Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.425534 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f","Type":"ContainerDied","Data":"418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9"} Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.425575 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f","Type":"ContainerDied","Data":"72aa688aa0526d362d38494bd9017044b31f9980b7efc632931311497de36836"} Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.425628 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.443069 4758 scope.go:117] "RemoveContainer" containerID="7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5" Oct 04 11:10:26 crc kubenswrapper[4758]: E1004 11:10:26.443548 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5\": container with ID starting with 7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5 not found: ID does not exist" containerID="7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.443589 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5"} err="failed to get container status \"7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5\": rpc error: code = NotFound desc = could not find container \"7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5\": container with ID starting with 7a462a05aa1846b808fb7596c1befe56cf2e308fb1aa0670c281e1f9ae1996c5 not found: ID does not exist" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.443613 4758 scope.go:117] "RemoveContainer" containerID="418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.446532 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-combined-ca-bundle\") pod \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.446686 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-config-data\") pod \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.446795 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-logs\") pod \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.446860 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxgr2\" (UniqueName: \"kubernetes.io/projected/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-kube-api-access-pxgr2\") pod \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\" (UID: \"d89c85d7-affe-4ba2-a61d-ea28ea0ec11f\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.447345 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-logs" (OuterVolumeSpecName: "logs") pod "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" (UID: "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.448503 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.451668 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-kube-api-access-pxgr2" (OuterVolumeSpecName: "kube-api-access-pxgr2") pod "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" (UID: "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f"). InnerVolumeSpecName "kube-api-access-pxgr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.465508 4758 scope.go:117] "RemoveContainer" containerID="0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.473459 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-config-data" (OuterVolumeSpecName: "config-data") pod "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" (UID: "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.476700 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" (UID: "d89c85d7-affe-4ba2-a61d-ea28ea0ec11f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.482917 4758 scope.go:117] "RemoveContainer" containerID="418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9" Oct 04 11:10:26 crc kubenswrapper[4758]: E1004 11:10:26.483313 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9\": container with ID starting with 418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9 not found: ID does not exist" containerID="418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.483351 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9"} err="failed to get container status \"418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9\": rpc error: code = NotFound desc = could not find container \"418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9\": container with ID starting with 418e8e28baeeccbacadcfc9e732d0e67ce19f16891b73f9eb9d3cfe0038e8bc9 not found: ID does not exist" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.483376 4758 scope.go:117] "RemoveContainer" containerID="0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5" Oct 04 11:10:26 crc kubenswrapper[4758]: E1004 11:10:26.483695 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5\": container with ID starting with 0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5 not found: ID does not exist" containerID="0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.483725 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5"} err="failed to get container status \"0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5\": rpc error: code = NotFound desc = could not find container \"0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5\": container with ID starting with 0ec0b3187487b810a3ef2320d476d5c15dca74c84b3a7c9a3c73da8356a9dbd5 not found: ID does not exist" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.549792 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-combined-ca-bundle\") pod \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.550285 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-config-data\") pod \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.550324 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7gmm\" (UniqueName: \"kubernetes.io/projected/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-kube-api-access-h7gmm\") pod \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\" (UID: \"3fff34a3-8666-465c-b9bd-6bd6e56c17fe\") " Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.551264 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.551904 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxgr2\" (UniqueName: \"kubernetes.io/projected/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-kube-api-access-pxgr2\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.551941 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.553225 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-kube-api-access-h7gmm" (OuterVolumeSpecName: "kube-api-access-h7gmm") pod "3fff34a3-8666-465c-b9bd-6bd6e56c17fe" (UID: "3fff34a3-8666-465c-b9bd-6bd6e56c17fe"). InnerVolumeSpecName "kube-api-access-h7gmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.572052 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fff34a3-8666-465c-b9bd-6bd6e56c17fe" (UID: "3fff34a3-8666-465c-b9bd-6bd6e56c17fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.584019 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-config-data" (OuterVolumeSpecName: "config-data") pod "3fff34a3-8666-465c-b9bd-6bd6e56c17fe" (UID: "3fff34a3-8666-465c-b9bd-6bd6e56c17fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.653184 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7gmm\" (UniqueName: \"kubernetes.io/projected/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-kube-api-access-h7gmm\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.653222 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.653234 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fff34a3-8666-465c-b9bd-6bd6e56c17fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.755227 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.761318 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.769959 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.779474 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.797399 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: E1004 11:10:26.797869 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-metadata" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.797927 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-metadata" Oct 04 11:10:26 crc kubenswrapper[4758]: E1004 11:10:26.797948 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-log" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.797956 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-log" Oct 04 11:10:26 crc kubenswrapper[4758]: E1004 11:10:26.798013 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fff34a3-8666-465c-b9bd-6bd6e56c17fe" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.798023 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fff34a3-8666-465c-b9bd-6bd6e56c17fe" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.798326 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fff34a3-8666-465c-b9bd-6bd6e56c17fe" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.798370 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-log" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.798386 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" containerName="nova-metadata-metadata" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.800236 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.807004 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.807264 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.808125 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.810000 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.817033 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.817226 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.817335 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.844830 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.853952 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880077 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880148 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880174 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lllgg\" (UniqueName: \"kubernetes.io/projected/f1907043-d54d-4502-bc9a-008711ae3ce1-kube-api-access-lllgg\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880238 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880268 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880285 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bz8p\" (UniqueName: \"kubernetes.io/projected/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-kube-api-access-7bz8p\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907043-d54d-4502-bc9a-008711ae3ce1-logs\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880352 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880377 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.880392 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-config-data\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983274 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983362 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983391 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bz8p\" (UniqueName: \"kubernetes.io/projected/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-kube-api-access-7bz8p\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983448 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907043-d54d-4502-bc9a-008711ae3ce1-logs\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983490 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983526 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983546 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-config-data\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983618 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983640 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.983667 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lllgg\" (UniqueName: \"kubernetes.io/projected/f1907043-d54d-4502-bc9a-008711ae3ce1-kube-api-access-lllgg\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.985331 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907043-d54d-4502-bc9a-008711ae3ce1-logs\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.988950 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-config-data\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.990168 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.990255 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.991088 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.991640 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.992159 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.997952 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:26 crc kubenswrapper[4758]: I1004 11:10:26.999542 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lllgg\" (UniqueName: \"kubernetes.io/projected/f1907043-d54d-4502-bc9a-008711ae3ce1-kube-api-access-lllgg\") pod \"nova-metadata-0\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " pod="openstack/nova-metadata-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.001192 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bz8p\" (UniqueName: \"kubernetes.io/projected/69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033-kube-api-access-7bz8p\") pod \"nova-cell1-novncproxy-0\" (UID: \"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.126547 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.143353 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.337209 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fff34a3-8666-465c-b9bd-6bd6e56c17fe" path="/var/lib/kubelet/pods/3fff34a3-8666-465c-b9bd-6bd6e56c17fe/volumes" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.338432 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d89c85d7-affe-4ba2-a61d-ea28ea0ec11f" path="/var/lib/kubelet/pods/d89c85d7-affe-4ba2-a61d-ea28ea0ec11f/volumes" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.608926 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 11:10:27 crc kubenswrapper[4758]: W1004 11:10:27.609494 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69b3b6d4_8c4d_4dd6_8836_eaf8a4a92033.slice/crio-0892aee1e89d38d73ac32e0dfe776907e3ceb39674e7277960b59c0c84ee0580 WatchSource:0}: Error finding container 0892aee1e89d38d73ac32e0dfe776907e3ceb39674e7277960b59c0c84ee0580: Status 404 returned error can't find the container with id 0892aee1e89d38d73ac32e0dfe776907e3ceb39674e7277960b59c0c84ee0580 Oct 04 11:10:27 crc kubenswrapper[4758]: W1004 11:10:27.623926 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1907043_d54d_4502_bc9a_008711ae3ce1.slice/crio-a3f612dd104dcb2eb8633f1df2e1e0e24ea20a143b3756c866a37baadd511600 WatchSource:0}: Error finding container a3f612dd104dcb2eb8633f1df2e1e0e24ea20a143b3756c866a37baadd511600: Status 404 returned error can't find the container with id a3f612dd104dcb2eb8633f1df2e1e0e24ea20a143b3756c866a37baadd511600 Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.624189 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.865219 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.865729 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.867878 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 11:10:27 crc kubenswrapper[4758]: I1004 11:10:27.869581 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.445894 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1907043-d54d-4502-bc9a-008711ae3ce1","Type":"ContainerStarted","Data":"bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd"} Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.445933 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1907043-d54d-4502-bc9a-008711ae3ce1","Type":"ContainerStarted","Data":"2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25"} Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.445943 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1907043-d54d-4502-bc9a-008711ae3ce1","Type":"ContainerStarted","Data":"a3f612dd104dcb2eb8633f1df2e1e0e24ea20a143b3756c866a37baadd511600"} Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.448454 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033","Type":"ContainerStarted","Data":"e5d3b6cedb9d4cfe7767998c169a7da115f9372869cbeae42b89e7448848b747"} Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.448493 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033","Type":"ContainerStarted","Data":"0892aee1e89d38d73ac32e0dfe776907e3ceb39674e7277960b59c0c84ee0580"} Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.448832 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.457446 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.467456 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.467433234 podStartE2EDuration="2.467433234s" podCreationTimestamp="2025-10-04 11:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:28.462553952 +0000 UTC m=+1205.755204851" watchObservedRunningTime="2025-10-04 11:10:28.467433234 +0000 UTC m=+1205.760084113" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.516989 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.516972692 podStartE2EDuration="2.516972692s" podCreationTimestamp="2025-10-04 11:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:28.508911514 +0000 UTC m=+1205.801562443" watchObservedRunningTime="2025-10-04 11:10:28.516972692 +0000 UTC m=+1205.809623581" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.727300 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-xbhdg"] Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.728716 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.796997 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-xbhdg"] Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.820030 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.820089 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.820184 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.820224 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ntl4\" (UniqueName: \"kubernetes.io/projected/ad4df944-5c12-46c7-8791-fc7aa5085e23-kube-api-access-9ntl4\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.820257 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.820297 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-config\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.932371 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ntl4\" (UniqueName: \"kubernetes.io/projected/ad4df944-5c12-46c7-8791-fc7aa5085e23-kube-api-access-9ntl4\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.932443 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.932502 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-config\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.932559 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.932593 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.932692 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.933500 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.934256 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-config\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.934597 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.941647 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.943785 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:28 crc kubenswrapper[4758]: I1004 11:10:28.981049 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ntl4\" (UniqueName: \"kubernetes.io/projected/ad4df944-5c12-46c7-8791-fc7aa5085e23-kube-api-access-9ntl4\") pod \"dnsmasq-dns-89c5cd4d5-xbhdg\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:29 crc kubenswrapper[4758]: I1004 11:10:29.045015 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:29 crc kubenswrapper[4758]: I1004 11:10:29.686503 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-xbhdg"] Oct 04 11:10:30 crc kubenswrapper[4758]: I1004 11:10:30.467705 4758 generic.go:334] "Generic (PLEG): container finished" podID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerID="007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53" exitCode=0 Oct 04 11:10:30 crc kubenswrapper[4758]: I1004 11:10:30.467799 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" event={"ID":"ad4df944-5c12-46c7-8791-fc7aa5085e23","Type":"ContainerDied","Data":"007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53"} Oct 04 11:10:30 crc kubenswrapper[4758]: I1004 11:10:30.468124 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" event={"ID":"ad4df944-5c12-46c7-8791-fc7aa5085e23","Type":"ContainerStarted","Data":"fb23bd6268e80256568597d79594330de37c4262f00f3750da3209fe61e62d8f"} Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.248992 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.249396 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.323531 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.323798 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-central-agent" containerID="cri-o://4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c" gracePeriod=30 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.323912 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="sg-core" containerID="cri-o://a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581" gracePeriod=30 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.323941 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="proxy-httpd" containerID="cri-o://a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc" gracePeriod=30 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.324092 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-notification-agent" containerID="cri-o://55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc" gracePeriod=30 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.433370 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.190:3000/\": read tcp 10.217.0.2:60132->10.217.0.190:3000: read: connection reset by peer" Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.476146 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.481685 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerID="a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581" exitCode=2 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.481740 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerDied","Data":"a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581"} Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.485268 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" event={"ID":"ad4df944-5c12-46c7-8791-fc7aa5085e23","Type":"ContainerStarted","Data":"cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658"} Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.485378 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-log" containerID="cri-o://9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319" gracePeriod=30 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.485461 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-api" containerID="cri-o://5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124" gracePeriod=30 Oct 04 11:10:31 crc kubenswrapper[4758]: I1004 11:10:31.506934 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" podStartSLOduration=3.506917994 podStartE2EDuration="3.506917994s" podCreationTimestamp="2025-10-04 11:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:31.505900797 +0000 UTC m=+1208.798551696" watchObservedRunningTime="2025-10-04 11:10:31.506917994 +0000 UTC m=+1208.799568873" Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.126997 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.127953 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.144624 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.497687 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerID="9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319" exitCode=143 Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.497768 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a6ab03a-ce43-4cf7-a03b-716538b8b26e","Type":"ContainerDied","Data":"9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319"} Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.501156 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerID="a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc" exitCode=0 Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.501193 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerID="4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c" exitCode=0 Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.501236 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerDied","Data":"a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc"} Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.501277 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerDied","Data":"4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c"} Oct 04 11:10:32 crc kubenswrapper[4758]: I1004 11:10:32.501587 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.019813 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.118600 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-scripts\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.118811 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-config-data\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.118883 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-log-httpd\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.118906 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-ceilometer-tls-certs\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.118936 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-sg-core-conf-yaml\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.121162 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-combined-ca-bundle\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.121294 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-run-httpd\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.121357 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqmgm\" (UniqueName: \"kubernetes.io/projected/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-kube-api-access-tqmgm\") pod \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\" (UID: \"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae\") " Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.122513 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.123753 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.134426 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-kube-api-access-tqmgm" (OuterVolumeSpecName: "kube-api-access-tqmgm") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "kube-api-access-tqmgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.144599 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-scripts" (OuterVolumeSpecName: "scripts") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.211241 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.225590 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.225703 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.225763 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.225829 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqmgm\" (UniqueName: \"kubernetes.io/projected/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-kube-api-access-tqmgm\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.225885 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.237743 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.244337 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.280527 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-config-data" (OuterVolumeSpecName: "config-data") pod "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" (UID: "c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.336809 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.336864 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.336886 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.517445 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerID="55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc" exitCode=0 Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.518440 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.519324 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerDied","Data":"55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc"} Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.519355 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae","Type":"ContainerDied","Data":"58282a3b14d402b604ec6ed41d52729cb15dfec963b08909166fbc20b200614b"} Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.519371 4758 scope.go:117] "RemoveContainer" containerID="a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.544919 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.548808 4758 scope.go:117] "RemoveContainer" containerID="a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.553460 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.565426 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.565787 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-notification-agent" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.565803 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-notification-agent" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.565811 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-central-agent" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.565819 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-central-agent" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.565828 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="sg-core" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.565833 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="sg-core" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.565850 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="proxy-httpd" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.565855 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="proxy-httpd" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.566015 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-notification-agent" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.566029 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="proxy-httpd" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.566044 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="sg-core" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.566054 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" containerName="ceilometer-central-agent" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.569474 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.574695 4758 scope.go:117] "RemoveContainer" containerID="55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.575436 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.575627 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.575740 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.608242 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.613197 4758 scope.go:117] "RemoveContainer" containerID="4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.633542 4758 scope.go:117] "RemoveContainer" containerID="a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.634201 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc\": container with ID starting with a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc not found: ID does not exist" containerID="a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.634240 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc"} err="failed to get container status \"a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc\": rpc error: code = NotFound desc = could not find container \"a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc\": container with ID starting with a814bec062feeaaa798f3685de6f859e801d8e6510100c9530276ab2bbbdb7fc not found: ID does not exist" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.634263 4758 scope.go:117] "RemoveContainer" containerID="a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.634637 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581\": container with ID starting with a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581 not found: ID does not exist" containerID="a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.634667 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581"} err="failed to get container status \"a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581\": rpc error: code = NotFound desc = could not find container \"a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581\": container with ID starting with a3b225699d0bc320d0349e4368014548d09c3d77b5bc07d25cd48d16d5bca581 not found: ID does not exist" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.634688 4758 scope.go:117] "RemoveContainer" containerID="55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.634949 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc\": container with ID starting with 55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc not found: ID does not exist" containerID="55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.634971 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc"} err="failed to get container status \"55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc\": rpc error: code = NotFound desc = could not find container \"55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc\": container with ID starting with 55833d76385b6bcf275adadd2b1041d8cd5e11a973f6768c3943a828b55a40cc not found: ID does not exist" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.634985 4758 scope.go:117] "RemoveContainer" containerID="4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c" Oct 04 11:10:33 crc kubenswrapper[4758]: E1004 11:10:33.635223 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c\": container with ID starting with 4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c not found: ID does not exist" containerID="4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.635246 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c"} err="failed to get container status \"4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c\": rpc error: code = NotFound desc = could not find container \"4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c\": container with ID starting with 4ac668012137b4508125f1df45d40f3f69846f60ddbeadb7ba6a9b0c2d21188c not found: ID does not exist" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643334 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643375 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643396 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln6dk\" (UniqueName: \"kubernetes.io/projected/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-kube-api-access-ln6dk\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643414 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643428 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-scripts\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643443 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-run-httpd\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643470 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-log-httpd\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.643511 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-config-data\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.744920 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.744974 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.744998 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln6dk\" (UniqueName: \"kubernetes.io/projected/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-kube-api-access-ln6dk\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745019 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745033 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-scripts\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745050 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-run-httpd\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745077 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-log-httpd\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745138 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-config-data\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745876 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-run-httpd\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.745893 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-log-httpd\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.749002 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-config-data\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.749574 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.751095 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.751498 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.759942 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-scripts\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.763264 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln6dk\" (UniqueName: \"kubernetes.io/projected/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-kube-api-access-ln6dk\") pod \"ceilometer-0\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " pod="openstack/ceilometer-0" Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.845085 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:33 crc kubenswrapper[4758]: I1004 11:10:33.846562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:34 crc kubenswrapper[4758]: I1004 11:10:34.349554 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:34 crc kubenswrapper[4758]: W1004 11:10:34.354043 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb1b97c2_b0d2_46ff_a06e_65436fbc02fa.slice/crio-52c05123647ea14edb5e2a564a9d5641662ee26a004d2e490773e0579033d7f3 WatchSource:0}: Error finding container 52c05123647ea14edb5e2a564a9d5641662ee26a004d2e490773e0579033d7f3: Status 404 returned error can't find the container with id 52c05123647ea14edb5e2a564a9d5641662ee26a004d2e490773e0579033d7f3 Oct 04 11:10:34 crc kubenswrapper[4758]: I1004 11:10:34.359574 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:10:34 crc kubenswrapper[4758]: I1004 11:10:34.529907 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerStarted","Data":"52c05123647ea14edb5e2a564a9d5641662ee26a004d2e490773e0579033d7f3"} Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.265559 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.335061 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae" path="/var/lib/kubelet/pods/c8f83afb-3b5f-4e5b-a29d-ba85af2ca4ae/volumes" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.373344 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-config-data\") pod \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.373417 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-logs\") pod \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.373537 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-combined-ca-bundle\") pod \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.373584 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9tq2\" (UniqueName: \"kubernetes.io/projected/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-kube-api-access-z9tq2\") pod \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\" (UID: \"7a6ab03a-ce43-4cf7-a03b-716538b8b26e\") " Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.379438 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-logs" (OuterVolumeSpecName: "logs") pod "7a6ab03a-ce43-4cf7-a03b-716538b8b26e" (UID: "7a6ab03a-ce43-4cf7-a03b-716538b8b26e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.379732 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-kube-api-access-z9tq2" (OuterVolumeSpecName: "kube-api-access-z9tq2") pod "7a6ab03a-ce43-4cf7-a03b-716538b8b26e" (UID: "7a6ab03a-ce43-4cf7-a03b-716538b8b26e"). InnerVolumeSpecName "kube-api-access-z9tq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.424296 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-config-data" (OuterVolumeSpecName: "config-data") pod "7a6ab03a-ce43-4cf7-a03b-716538b8b26e" (UID: "7a6ab03a-ce43-4cf7-a03b-716538b8b26e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.446040 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a6ab03a-ce43-4cf7-a03b-716538b8b26e" (UID: "7a6ab03a-ce43-4cf7-a03b-716538b8b26e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.476938 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.476969 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.476979 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.476987 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9tq2\" (UniqueName: \"kubernetes.io/projected/7a6ab03a-ce43-4cf7-a03b-716538b8b26e-kube-api-access-z9tq2\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.577764 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerStarted","Data":"ab66472cb012632582149773bb0ba4d43c2071783c849d63852be102f7970d6d"} Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.641321 4758 generic.go:334] "Generic (PLEG): container finished" podID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerID="5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124" exitCode=0 Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.641367 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a6ab03a-ce43-4cf7-a03b-716538b8b26e","Type":"ContainerDied","Data":"5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124"} Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.641404 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a6ab03a-ce43-4cf7-a03b-716538b8b26e","Type":"ContainerDied","Data":"364b97701067343c7e792fef4e6b3deb8910954cd5486f2a221384a8e56c0a38"} Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.641418 4758 scope.go:117] "RemoveContainer" containerID="5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.641544 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.741269 4758 scope.go:117] "RemoveContainer" containerID="9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.742912 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.778832 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.807574 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:35 crc kubenswrapper[4758]: E1004 11:10:35.808268 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-api" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.808285 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-api" Oct 04 11:10:35 crc kubenswrapper[4758]: E1004 11:10:35.808311 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-log" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.808317 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-log" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.808493 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-log" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.808524 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" containerName="nova-api-api" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.809616 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.815688 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.815868 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.815879 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.833626 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.843546 4758 scope.go:117] "RemoveContainer" containerID="5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124" Oct 04 11:10:35 crc kubenswrapper[4758]: E1004 11:10:35.846223 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124\": container with ID starting with 5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124 not found: ID does not exist" containerID="5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.846265 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124"} err="failed to get container status \"5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124\": rpc error: code = NotFound desc = could not find container \"5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124\": container with ID starting with 5f7a551e59c85111b87f6b16d46fe99058524c291ace471619c63992f4802124 not found: ID does not exist" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.846289 4758 scope.go:117] "RemoveContainer" containerID="9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319" Oct 04 11:10:35 crc kubenswrapper[4758]: E1004 11:10:35.846755 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319\": container with ID starting with 9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319 not found: ID does not exist" containerID="9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.846779 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319"} err="failed to get container status \"9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319\": rpc error: code = NotFound desc = could not find container \"9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319\": container with ID starting with 9d97d422617d71ef319fd39e910a53a7f82cf52f0f4cbd1604eabc3133063319 not found: ID does not exist" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.986659 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a89218c-9de3-466f-9fa5-c6c0cce8053d-logs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.986731 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.986815 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.986936 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85bkj\" (UniqueName: \"kubernetes.io/projected/3a89218c-9de3-466f-9fa5-c6c0cce8053d-kube-api-access-85bkj\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.987181 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-config-data\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:35 crc kubenswrapper[4758]: I1004 11:10:35.987240 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-public-tls-certs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089005 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a89218c-9de3-466f-9fa5-c6c0cce8053d-logs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089053 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089092 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85bkj\" (UniqueName: \"kubernetes.io/projected/3a89218c-9de3-466f-9fa5-c6c0cce8053d-kube-api-access-85bkj\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089202 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-config-data\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089222 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-public-tls-certs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.089456 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a89218c-9de3-466f-9fa5-c6c0cce8053d-logs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.093376 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.094123 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-config-data\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.094421 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-public-tls-certs\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.098957 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.113641 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85bkj\" (UniqueName: \"kubernetes.io/projected/3a89218c-9de3-466f-9fa5-c6c0cce8053d-kube-api-access-85bkj\") pod \"nova-api-0\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.161562 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.654822 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerStarted","Data":"0268f458691fc1591422ad2287f78c438633e16b6809d0308b7e5a22371f6e3e"} Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.655273 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerStarted","Data":"4f7b685973131ab19acb6d0dcf7c88a1837cc42295811ae53ac1e20a5296641a"} Oct 04 11:10:36 crc kubenswrapper[4758]: W1004 11:10:36.660425 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a89218c_9de3_466f_9fa5_c6c0cce8053d.slice/crio-6d90478a495aacfb146109a48266a4184aa3d2143d5e1eeb6669895fa87be7dc WatchSource:0}: Error finding container 6d90478a495aacfb146109a48266a4184aa3d2143d5e1eeb6669895fa87be7dc: Status 404 returned error can't find the container with id 6d90478a495aacfb146109a48266a4184aa3d2143d5e1eeb6669895fa87be7dc Oct 04 11:10:36 crc kubenswrapper[4758]: I1004 11:10:36.665770 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.126776 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.128801 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.144091 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.165141 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.336714 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6ab03a-ce43-4cf7-a03b-716538b8b26e" path="/var/lib/kubelet/pods/7a6ab03a-ce43-4cf7-a03b-716538b8b26e/volumes" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.680530 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a89218c-9de3-466f-9fa5-c6c0cce8053d","Type":"ContainerStarted","Data":"f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877"} Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.680562 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a89218c-9de3-466f-9fa5-c6c0cce8053d","Type":"ContainerStarted","Data":"05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0"} Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.680571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a89218c-9de3-466f-9fa5-c6c0cce8053d","Type":"ContainerStarted","Data":"6d90478a495aacfb146109a48266a4184aa3d2143d5e1eeb6669895fa87be7dc"} Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.699609 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.699590032 podStartE2EDuration="2.699590032s" podCreationTimestamp="2025-10-04 11:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:37.694985258 +0000 UTC m=+1214.987636147" watchObservedRunningTime="2025-10-04 11:10:37.699590032 +0000 UTC m=+1214.992240921" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.711784 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.904569 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-kw82l"] Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.906083 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.911701 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.911945 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 11:10:37 crc kubenswrapper[4758]: I1004 11:10:37.919503 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kw82l"] Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.029667 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzt5t\" (UniqueName: \"kubernetes.io/projected/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-kube-api-access-xzt5t\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.029749 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.029811 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-config-data\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.029835 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-scripts\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.131288 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-config-data\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.131359 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-scripts\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.131452 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzt5t\" (UniqueName: \"kubernetes.io/projected/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-kube-api-access-xzt5t\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.131536 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.133308 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.133342 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.136054 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.137679 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-config-data\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.142543 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-scripts\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.159232 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzt5t\" (UniqueName: \"kubernetes.io/projected/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-kube-api-access-xzt5t\") pod \"nova-cell1-cell-mapping-kw82l\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.240853 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.693042 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-central-agent" containerID="cri-o://ab66472cb012632582149773bb0ba4d43c2071783c849d63852be102f7970d6d" gracePeriod=30 Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.693430 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerStarted","Data":"ec30e85dedfaa2f6f948ef75104b0fffae7df9b6e1d1039bd6dfb1393006349c"} Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.693552 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="proxy-httpd" containerID="cri-o://ec30e85dedfaa2f6f948ef75104b0fffae7df9b6e1d1039bd6dfb1393006349c" gracePeriod=30 Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.693688 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="sg-core" containerID="cri-o://0268f458691fc1591422ad2287f78c438633e16b6809d0308b7e5a22371f6e3e" gracePeriod=30 Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.693723 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-notification-agent" containerID="cri-o://4f7b685973131ab19acb6d0dcf7c88a1837cc42295811ae53ac1e20a5296641a" gracePeriod=30 Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.693990 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.730536 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4567777619999998 podStartE2EDuration="5.730521555s" podCreationTimestamp="2025-10-04 11:10:33 +0000 UTC" firstStartedPulling="2025-10-04 11:10:34.359340703 +0000 UTC m=+1211.651991592" lastFinishedPulling="2025-10-04 11:10:37.633084486 +0000 UTC m=+1214.925735385" observedRunningTime="2025-10-04 11:10:38.723975478 +0000 UTC m=+1216.016626367" watchObservedRunningTime="2025-10-04 11:10:38.730521555 +0000 UTC m=+1216.023172444" Oct 04 11:10:38 crc kubenswrapper[4758]: I1004 11:10:38.785999 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kw82l"] Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.046249 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.097776 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-c4qcg"] Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.102635 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="dnsmasq-dns" containerID="cri-o://8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83" gracePeriod=10 Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.643771 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.728434 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kw82l" event={"ID":"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b","Type":"ContainerStarted","Data":"7ccc35e0e3932fe89919b7a550264f5d3ffe2e6a8bf754440f7b9e225b157765"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.728480 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kw82l" event={"ID":"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b","Type":"ContainerStarted","Data":"39d6adb82bee664621301a0eec6adbd282a8bb746b2ada1a16775d5a78e00364"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.731185 4758 generic.go:334] "Generic (PLEG): container finished" podID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerID="ec30e85dedfaa2f6f948ef75104b0fffae7df9b6e1d1039bd6dfb1393006349c" exitCode=0 Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.731209 4758 generic.go:334] "Generic (PLEG): container finished" podID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerID="0268f458691fc1591422ad2287f78c438633e16b6809d0308b7e5a22371f6e3e" exitCode=2 Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.731216 4758 generic.go:334] "Generic (PLEG): container finished" podID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerID="4f7b685973131ab19acb6d0dcf7c88a1837cc42295811ae53ac1e20a5296641a" exitCode=0 Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.731247 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerDied","Data":"ec30e85dedfaa2f6f948ef75104b0fffae7df9b6e1d1039bd6dfb1393006349c"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.731262 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerDied","Data":"0268f458691fc1591422ad2287f78c438633e16b6809d0308b7e5a22371f6e3e"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.731272 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerDied","Data":"4f7b685973131ab19acb6d0dcf7c88a1837cc42295811ae53ac1e20a5296641a"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.732462 4758 generic.go:334] "Generic (PLEG): container finished" podID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerID="8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83" exitCode=0 Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.732487 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" event={"ID":"c710639c-0dc3-43ba-b084-fee7da532dc8","Type":"ContainerDied","Data":"8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.732502 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" event={"ID":"c710639c-0dc3-43ba-b084-fee7da532dc8","Type":"ContainerDied","Data":"34eb4f705bec27171cb22c31da010c9182772b85845a008a9acdf12f7b76d586"} Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.732518 4758 scope.go:117] "RemoveContainer" containerID="8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.732608 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.749390 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-kw82l" podStartSLOduration=2.749366281 podStartE2EDuration="2.749366281s" podCreationTimestamp="2025-10-04 11:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:39.749175136 +0000 UTC m=+1217.041826025" watchObservedRunningTime="2025-10-04 11:10:39.749366281 +0000 UTC m=+1217.042017180" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.781682 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-svc\") pod \"c710639c-0dc3-43ba-b084-fee7da532dc8\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.781754 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56gbv\" (UniqueName: \"kubernetes.io/projected/c710639c-0dc3-43ba-b084-fee7da532dc8-kube-api-access-56gbv\") pod \"c710639c-0dc3-43ba-b084-fee7da532dc8\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.781815 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-nb\") pod \"c710639c-0dc3-43ba-b084-fee7da532dc8\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.781923 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-sb\") pod \"c710639c-0dc3-43ba-b084-fee7da532dc8\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.782018 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-swift-storage-0\") pod \"c710639c-0dc3-43ba-b084-fee7da532dc8\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.782038 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-config\") pod \"c710639c-0dc3-43ba-b084-fee7da532dc8\" (UID: \"c710639c-0dc3-43ba-b084-fee7da532dc8\") " Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.782630 4758 scope.go:117] "RemoveContainer" containerID="cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.845780 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c710639c-0dc3-43ba-b084-fee7da532dc8-kube-api-access-56gbv" (OuterVolumeSpecName: "kube-api-access-56gbv") pod "c710639c-0dc3-43ba-b084-fee7da532dc8" (UID: "c710639c-0dc3-43ba-b084-fee7da532dc8"). InnerVolumeSpecName "kube-api-access-56gbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.861552 4758 scope.go:117] "RemoveContainer" containerID="8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83" Oct 04 11:10:39 crc kubenswrapper[4758]: E1004 11:10:39.876046 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83\": container with ID starting with 8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83 not found: ID does not exist" containerID="8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.876191 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83"} err="failed to get container status \"8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83\": rpc error: code = NotFound desc = could not find container \"8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83\": container with ID starting with 8c935b31e036cac0a98c516c40863aa36d5b48bd3ab05e1d1ce10cd558ca1c83 not found: ID does not exist" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.876273 4758 scope.go:117] "RemoveContainer" containerID="cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349" Oct 04 11:10:39 crc kubenswrapper[4758]: E1004 11:10:39.880300 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349\": container with ID starting with cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349 not found: ID does not exist" containerID="cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.880353 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349"} err="failed to get container status \"cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349\": rpc error: code = NotFound desc = could not find container \"cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349\": container with ID starting with cb2c33bf5b30af37f2a232b741ca8affb2475d854bfc75571fd42e9d18c42349 not found: ID does not exist" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.886225 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56gbv\" (UniqueName: \"kubernetes.io/projected/c710639c-0dc3-43ba-b084-fee7da532dc8-kube-api-access-56gbv\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.970458 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c710639c-0dc3-43ba-b084-fee7da532dc8" (UID: "c710639c-0dc3-43ba-b084-fee7da532dc8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:39 crc kubenswrapper[4758]: I1004 11:10:39.993030 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.026533 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-config" (OuterVolumeSpecName: "config") pod "c710639c-0dc3-43ba-b084-fee7da532dc8" (UID: "c710639c-0dc3-43ba-b084-fee7da532dc8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.040538 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c710639c-0dc3-43ba-b084-fee7da532dc8" (UID: "c710639c-0dc3-43ba-b084-fee7da532dc8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.042250 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c710639c-0dc3-43ba-b084-fee7da532dc8" (UID: "c710639c-0dc3-43ba-b084-fee7da532dc8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.044658 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c710639c-0dc3-43ba-b084-fee7da532dc8" (UID: "c710639c-0dc3-43ba-b084-fee7da532dc8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.094990 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.095031 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.095042 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.095051 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c710639c-0dc3-43ba-b084-fee7da532dc8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.361345 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-c4qcg"] Oct 04 11:10:40 crc kubenswrapper[4758]: I1004 11:10:40.369019 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-c4qcg"] Oct 04 11:10:41 crc kubenswrapper[4758]: I1004 11:10:41.335930 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" path="/var/lib/kubelet/pods/c710639c-0dc3-43ba-b084-fee7da532dc8/volumes" Oct 04 11:10:42 crc kubenswrapper[4758]: I1004 11:10:42.761608 4758 generic.go:334] "Generic (PLEG): container finished" podID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerID="ab66472cb012632582149773bb0ba4d43c2071783c849d63852be102f7970d6d" exitCode=0 Oct 04 11:10:42 crc kubenswrapper[4758]: I1004 11:10:42.761648 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerDied","Data":"ab66472cb012632582149773bb0ba4d43c2071783c849d63852be102f7970d6d"} Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.053955 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246009 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-ceilometer-tls-certs\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246277 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-config-data\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246359 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-combined-ca-bundle\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246453 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-sg-core-conf-yaml\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246540 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln6dk\" (UniqueName: \"kubernetes.io/projected/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-kube-api-access-ln6dk\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246681 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-scripts\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246749 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-run-httpd\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.246833 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-log-httpd\") pod \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\" (UID: \"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa\") " Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.247555 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.248517 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.254108 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-scripts" (OuterVolumeSpecName: "scripts") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.254125 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-kube-api-access-ln6dk" (OuterVolumeSpecName: "kube-api-access-ln6dk") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "kube-api-access-ln6dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.278816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.312797 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.319810 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350679 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350874 4758 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350886 4758 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350894 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350908 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350916 4758 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.350926 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln6dk\" (UniqueName: \"kubernetes.io/projected/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-kube-api-access-ln6dk\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.375352 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-config-data" (OuterVolumeSpecName: "config-data") pod "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" (UID: "eb1b97c2-b0d2-46ff-a06e-65436fbc02fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.452332 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.774163 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb1b97c2-b0d2-46ff-a06e-65436fbc02fa","Type":"ContainerDied","Data":"52c05123647ea14edb5e2a564a9d5641662ee26a004d2e490773e0579033d7f3"} Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.774230 4758 scope.go:117] "RemoveContainer" containerID="ec30e85dedfaa2f6f948ef75104b0fffae7df9b6e1d1039bd6dfb1393006349c" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.774310 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.831349 4758 scope.go:117] "RemoveContainer" containerID="0268f458691fc1591422ad2287f78c438633e16b6809d0308b7e5a22371f6e3e" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.846696 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.855750 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.876297 4758 scope.go:117] "RemoveContainer" containerID="4f7b685973131ab19acb6d0dcf7c88a1837cc42295811ae53ac1e20a5296641a" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.881763 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:43 crc kubenswrapper[4758]: E1004 11:10:43.882140 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="proxy-httpd" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882151 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="proxy-httpd" Oct 04 11:10:43 crc kubenswrapper[4758]: E1004 11:10:43.882165 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="init" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882171 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="init" Oct 04 11:10:43 crc kubenswrapper[4758]: E1004 11:10:43.882179 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-notification-agent" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882184 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-notification-agent" Oct 04 11:10:43 crc kubenswrapper[4758]: E1004 11:10:43.882205 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="sg-core" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882211 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="sg-core" Oct 04 11:10:43 crc kubenswrapper[4758]: E1004 11:10:43.882220 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="dnsmasq-dns" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882225 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="dnsmasq-dns" Oct 04 11:10:43 crc kubenswrapper[4758]: E1004 11:10:43.882238 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-central-agent" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882244 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-central-agent" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882422 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="dnsmasq-dns" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882440 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="proxy-httpd" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882449 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-notification-agent" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882457 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="ceilometer-central-agent" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.882473 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" containerName="sg-core" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.896967 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.899859 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.899968 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.908181 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.915311 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:43 crc kubenswrapper[4758]: I1004 11:10:43.934159 4758 scope.go:117] "RemoveContainer" containerID="ab66472cb012632582149773bb0ba4d43c2071783c849d63852be102f7970d6d" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063497 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063550 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-scripts\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063599 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063632 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68w95\" (UniqueName: \"kubernetes.io/projected/9e168ea3-6a31-4a16-b478-7b25e649810e-kube-api-access-68w95\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063650 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e168ea3-6a31-4a16-b478-7b25e649810e-log-httpd\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063671 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-config-data\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063716 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e168ea3-6a31-4a16-b478-7b25e649810e-run-httpd\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.063732 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.164941 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165002 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68w95\" (UniqueName: \"kubernetes.io/projected/9e168ea3-6a31-4a16-b478-7b25e649810e-kube-api-access-68w95\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165025 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e168ea3-6a31-4a16-b478-7b25e649810e-log-httpd\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165052 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-config-data\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e168ea3-6a31-4a16-b478-7b25e649810e-run-httpd\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165164 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165204 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165230 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-scripts\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165572 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e168ea3-6a31-4a16-b478-7b25e649810e-log-httpd\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.165576 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9e168ea3-6a31-4a16-b478-7b25e649810e-run-httpd\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.170792 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-scripts\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.171310 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.176893 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-config-data\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.177780 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.181301 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68w95\" (UniqueName: \"kubernetes.io/projected/9e168ea3-6a31-4a16-b478-7b25e649810e-kube-api-access-68w95\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.184417 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9e168ea3-6a31-4a16-b478-7b25e649810e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9e168ea3-6a31-4a16-b478-7b25e649810e\") " pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.223610 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.496843 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.591683 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-c4qcg" podUID="c710639c-0dc3-43ba-b084-fee7da532dc8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.184:5353: i/o timeout" Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.785599 4758 generic.go:334] "Generic (PLEG): container finished" podID="3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" containerID="7ccc35e0e3932fe89919b7a550264f5d3ffe2e6a8bf754440f7b9e225b157765" exitCode=0 Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.785663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kw82l" event={"ID":"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b","Type":"ContainerDied","Data":"7ccc35e0e3932fe89919b7a550264f5d3ffe2e6a8bf754440f7b9e225b157765"} Oct 04 11:10:44 crc kubenswrapper[4758]: I1004 11:10:44.789663 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e168ea3-6a31-4a16-b478-7b25e649810e","Type":"ContainerStarted","Data":"070d4df65f2460ab6244cef37f3ad7bd13ab3dcdb227de6e4f3480bf8abe1952"} Oct 04 11:10:45 crc kubenswrapper[4758]: I1004 11:10:45.341853 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1b97c2-b0d2-46ff-a06e-65436fbc02fa" path="/var/lib/kubelet/pods/eb1b97c2-b0d2-46ff-a06e-65436fbc02fa/volumes" Oct 04 11:10:45 crc kubenswrapper[4758]: I1004 11:10:45.821634 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e168ea3-6a31-4a16-b478-7b25e649810e","Type":"ContainerStarted","Data":"b56fd260e300f4b8456e3e1573cd1951ef6c35a43a4d120392b40ca3b1669041"} Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.163451 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.163953 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.194499 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.311690 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzt5t\" (UniqueName: \"kubernetes.io/projected/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-kube-api-access-xzt5t\") pod \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.311752 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-config-data\") pod \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.312452 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-scripts\") pod \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.312569 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-combined-ca-bundle\") pod \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\" (UID: \"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b\") " Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.327695 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-kube-api-access-xzt5t" (OuterVolumeSpecName: "kube-api-access-xzt5t") pod "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" (UID: "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b"). InnerVolumeSpecName "kube-api-access-xzt5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.338910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-scripts" (OuterVolumeSpecName: "scripts") pod "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" (UID: "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.355131 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" (UID: "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.364461 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-config-data" (OuterVolumeSpecName: "config-data") pod "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" (UID: "3c16e6f6-a6f4-49e4-b8e7-df0d923e801b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.414628 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzt5t\" (UniqueName: \"kubernetes.io/projected/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-kube-api-access-xzt5t\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.414893 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.414903 4758 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.414911 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.833194 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kw82l" event={"ID":"3c16e6f6-a6f4-49e4-b8e7-df0d923e801b","Type":"ContainerDied","Data":"39d6adb82bee664621301a0eec6adbd282a8bb746b2ada1a16775d5a78e00364"} Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.833231 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39d6adb82bee664621301a0eec6adbd282a8bb746b2ada1a16775d5a78e00364" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.833304 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kw82l" Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.855169 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e168ea3-6a31-4a16-b478-7b25e649810e","Type":"ContainerStarted","Data":"7f73dc9a0847d53a290fa689f426317d929bd67e36c767c5806e37b13d10cc3d"} Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.855206 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e168ea3-6a31-4a16-b478-7b25e649810e","Type":"ContainerStarted","Data":"0c6aeb5734ba3ad3aca672bc315063be1032973d8681317efa1cfa693b278c06"} Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.987478 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.998050 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:46 crc kubenswrapper[4758]: I1004 11:10:46.998276 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" containerName="nova-scheduler-scheduler" containerID="cri-o://d4a16210f529fcab2281e66777f877df4dccd8bc8b7a7fbf5d4a5d9d577f5136" gracePeriod=30 Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.027899 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.034598 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-log" containerID="cri-o://2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25" gracePeriod=30 Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.034698 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-metadata" containerID="cri-o://bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd" gracePeriod=30 Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.178248 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.178261 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.880572 4758 generic.go:334] "Generic (PLEG): container finished" podID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerID="2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25" exitCode=143 Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.880774 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-log" containerID="cri-o://05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0" gracePeriod=30 Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.880844 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1907043-d54d-4502-bc9a-008711ae3ce1","Type":"ContainerDied","Data":"2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25"} Oct 04 11:10:47 crc kubenswrapper[4758]: I1004 11:10:47.881187 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-api" containerID="cri-o://f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877" gracePeriod=30 Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.909008 4758 generic.go:334] "Generic (PLEG): container finished" podID="0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" containerID="d4a16210f529fcab2281e66777f877df4dccd8bc8b7a7fbf5d4a5d9d577f5136" exitCode=0 Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.909519 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460","Type":"ContainerDied","Data":"d4a16210f529fcab2281e66777f877df4dccd8bc8b7a7fbf5d4a5d9d577f5136"} Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.915638 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerID="05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0" exitCode=143 Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.915743 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a89218c-9de3-466f-9fa5-c6c0cce8053d","Type":"ContainerDied","Data":"05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0"} Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.946415 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9e168ea3-6a31-4a16-b478-7b25e649810e","Type":"ContainerStarted","Data":"90b512edbdc87b4bd0467f3e52c5b34f719ed7a66fb22486236ebc097b898fd9"} Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.946616 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.967534 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7549585690000002 podStartE2EDuration="5.967520151s" podCreationTimestamp="2025-10-04 11:10:43 +0000 UTC" firstStartedPulling="2025-10-04 11:10:44.506849612 +0000 UTC m=+1221.799500501" lastFinishedPulling="2025-10-04 11:10:47.719411194 +0000 UTC m=+1225.012062083" observedRunningTime="2025-10-04 11:10:48.962872325 +0000 UTC m=+1226.255523214" watchObservedRunningTime="2025-10-04 11:10:48.967520151 +0000 UTC m=+1226.260171040" Oct 04 11:10:48 crc kubenswrapper[4758]: I1004 11:10:48.995903 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.061949 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-config-data\") pod \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.061990 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-combined-ca-bundle\") pod \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.062036 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb4gb\" (UniqueName: \"kubernetes.io/projected/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-kube-api-access-kb4gb\") pod \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\" (UID: \"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460\") " Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.075331 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-kube-api-access-kb4gb" (OuterVolumeSpecName: "kube-api-access-kb4gb") pod "0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" (UID: "0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460"). InnerVolumeSpecName "kube-api-access-kb4gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.094309 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-config-data" (OuterVolumeSpecName: "config-data") pod "0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" (UID: "0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.107231 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" (UID: "0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.165084 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.165126 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.165137 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb4gb\" (UniqueName: \"kubernetes.io/projected/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460-kube-api-access-kb4gb\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.957862 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.958010 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460","Type":"ContainerDied","Data":"bdb3eefb404add15bb877407215b510afdd5b0ea58e1397fd3349bd9f0beab56"} Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.958229 4758 scope.go:117] "RemoveContainer" containerID="d4a16210f529fcab2281e66777f877df4dccd8bc8b7a7fbf5d4a5d9d577f5136" Oct 04 11:10:49 crc kubenswrapper[4758]: I1004 11:10:49.992621 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.012124 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.019899 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:50 crc kubenswrapper[4758]: E1004 11:10:50.020271 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" containerName="nova-manage" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.020288 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" containerName="nova-manage" Oct 04 11:10:50 crc kubenswrapper[4758]: E1004 11:10:50.020316 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" containerName="nova-scheduler-scheduler" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.020323 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" containerName="nova-scheduler-scheduler" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.020483 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" containerName="nova-scheduler-scheduler" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.020510 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" containerName="nova-manage" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.021070 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.040855 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.054851 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.091857 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e96b041-d262-4169-8ee6-cfee13404759-config-data\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.091926 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e96b041-d262-4169-8ee6-cfee13404759-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.092084 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5b5x\" (UniqueName: \"kubernetes.io/projected/0e96b041-d262-4169-8ee6-cfee13404759-kube-api-access-h5b5x\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.194114 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e96b041-d262-4169-8ee6-cfee13404759-config-data\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.194181 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e96b041-d262-4169-8ee6-cfee13404759-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.194275 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5b5x\" (UniqueName: \"kubernetes.io/projected/0e96b041-d262-4169-8ee6-cfee13404759-kube-api-access-h5b5x\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.201653 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e96b041-d262-4169-8ee6-cfee13404759-config-data\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.208869 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e96b041-d262-4169-8ee6-cfee13404759-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.213582 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5b5x\" (UniqueName: \"kubernetes.io/projected/0e96b041-d262-4169-8ee6-cfee13404759-kube-api-access-h5b5x\") pod \"nova-scheduler-0\" (UID: \"0e96b041-d262-4169-8ee6-cfee13404759\") " pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.358799 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.749531 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.808861 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lllgg\" (UniqueName: \"kubernetes.io/projected/f1907043-d54d-4502-bc9a-008711ae3ce1-kube-api-access-lllgg\") pod \"f1907043-d54d-4502-bc9a-008711ae3ce1\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.809446 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-config-data\") pod \"f1907043-d54d-4502-bc9a-008711ae3ce1\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.809511 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-nova-metadata-tls-certs\") pod \"f1907043-d54d-4502-bc9a-008711ae3ce1\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.809539 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-combined-ca-bundle\") pod \"f1907043-d54d-4502-bc9a-008711ae3ce1\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.809609 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907043-d54d-4502-bc9a-008711ae3ce1-logs\") pod \"f1907043-d54d-4502-bc9a-008711ae3ce1\" (UID: \"f1907043-d54d-4502-bc9a-008711ae3ce1\") " Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.811863 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1907043-d54d-4502-bc9a-008711ae3ce1-logs" (OuterVolumeSpecName: "logs") pod "f1907043-d54d-4502-bc9a-008711ae3ce1" (UID: "f1907043-d54d-4502-bc9a-008711ae3ce1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.818561 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1907043-d54d-4502-bc9a-008711ae3ce1-kube-api-access-lllgg" (OuterVolumeSpecName: "kube-api-access-lllgg") pod "f1907043-d54d-4502-bc9a-008711ae3ce1" (UID: "f1907043-d54d-4502-bc9a-008711ae3ce1"). InnerVolumeSpecName "kube-api-access-lllgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.845247 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-config-data" (OuterVolumeSpecName: "config-data") pod "f1907043-d54d-4502-bc9a-008711ae3ce1" (UID: "f1907043-d54d-4502-bc9a-008711ae3ce1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.852383 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1907043-d54d-4502-bc9a-008711ae3ce1" (UID: "f1907043-d54d-4502-bc9a-008711ae3ce1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.868728 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f1907043-d54d-4502-bc9a-008711ae3ce1" (UID: "f1907043-d54d-4502-bc9a-008711ae3ce1"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:50 crc kubenswrapper[4758]: W1004 11:10:50.907427 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e96b041_d262_4169_8ee6_cfee13404759.slice/crio-8fc6f528af6c61f335faba114c33cea5e0c36a2530d48889c7914b4f8e0e21f4 WatchSource:0}: Error finding container 8fc6f528af6c61f335faba114c33cea5e0c36a2530d48889c7914b4f8e0e21f4: Status 404 returned error can't find the container with id 8fc6f528af6c61f335faba114c33cea5e0c36a2530d48889c7914b4f8e0e21f4 Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.908606 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.911969 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.912005 4758 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.912021 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1907043-d54d-4502-bc9a-008711ae3ce1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.912033 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1907043-d54d-4502-bc9a-008711ae3ce1-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.912045 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lllgg\" (UniqueName: \"kubernetes.io/projected/f1907043-d54d-4502-bc9a-008711ae3ce1-kube-api-access-lllgg\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.992981 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0e96b041-d262-4169-8ee6-cfee13404759","Type":"ContainerStarted","Data":"8fc6f528af6c61f335faba114c33cea5e0c36a2530d48889c7914b4f8e0e21f4"} Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.995093 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.995121 4758 generic.go:334] "Generic (PLEG): container finished" podID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerID="bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd" exitCode=0 Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.995149 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1907043-d54d-4502-bc9a-008711ae3ce1","Type":"ContainerDied","Data":"bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd"} Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.995171 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1907043-d54d-4502-bc9a-008711ae3ce1","Type":"ContainerDied","Data":"a3f612dd104dcb2eb8633f1df2e1e0e24ea20a143b3756c866a37baadd511600"} Oct 04 11:10:50 crc kubenswrapper[4758]: I1004 11:10:50.995186 4758 scope.go:117] "RemoveContainer" containerID="bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.017842 4758 scope.go:117] "RemoveContainer" containerID="2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.051274 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.070919 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.075087 4758 scope.go:117] "RemoveContainer" containerID="bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd" Oct 04 11:10:51 crc kubenswrapper[4758]: E1004 11:10:51.075480 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd\": container with ID starting with bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd not found: ID does not exist" containerID="bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.075508 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd"} err="failed to get container status \"bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd\": rpc error: code = NotFound desc = could not find container \"bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd\": container with ID starting with bdbc16c81dac5b2064e3b193c2f14fd93f1dde4e2ea69e3e82ab6c3e66ff6dbd not found: ID does not exist" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.075526 4758 scope.go:117] "RemoveContainer" containerID="2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25" Oct 04 11:10:51 crc kubenswrapper[4758]: E1004 11:10:51.075922 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25\": container with ID starting with 2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25 not found: ID does not exist" containerID="2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.075944 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25"} err="failed to get container status \"2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25\": rpc error: code = NotFound desc = could not find container \"2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25\": container with ID starting with 2d5c101118be07ef75457671869d3903f95139b19a32b5a4e293ed7c0951db25 not found: ID does not exist" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.088354 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:51 crc kubenswrapper[4758]: E1004 11:10:51.088679 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-metadata" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.088697 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-metadata" Oct 04 11:10:51 crc kubenswrapper[4758]: E1004 11:10:51.088741 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-log" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.088749 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-log" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.088928 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-log" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.088941 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" containerName="nova-metadata-metadata" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.089929 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.092673 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.092829 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.096640 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.215461 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lmkr\" (UniqueName: \"kubernetes.io/projected/124e2bc8-911c-4842-be08-4674929ed4fb-kube-api-access-7lmkr\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.215710 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.215734 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-config-data\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.215756 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/124e2bc8-911c-4842-be08-4674929ed4fb-logs\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.215798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.317143 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lmkr\" (UniqueName: \"kubernetes.io/projected/124e2bc8-911c-4842-be08-4674929ed4fb-kube-api-access-7lmkr\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.317388 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.317477 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-config-data\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.317569 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/124e2bc8-911c-4842-be08-4674929ed4fb-logs\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.317686 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.318332 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/124e2bc8-911c-4842-be08-4674929ed4fb-logs\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.322304 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-config-data\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.322649 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.323923 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/124e2bc8-911c-4842-be08-4674929ed4fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.335937 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lmkr\" (UniqueName: \"kubernetes.io/projected/124e2bc8-911c-4842-be08-4674929ed4fb-kube-api-access-7lmkr\") pod \"nova-metadata-0\" (UID: \"124e2bc8-911c-4842-be08-4674929ed4fb\") " pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.338340 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460" path="/var/lib/kubelet/pods/0a73c1d4-ecc4-452c-85a1-c0f8d0fc8460/volumes" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.339223 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1907043-d54d-4502-bc9a-008711ae3ce1" path="/var/lib/kubelet/pods/f1907043-d54d-4502-bc9a-008711ae3ce1/volumes" Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.406483 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 11:10:51 crc kubenswrapper[4758]: W1004 11:10:51.966440 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod124e2bc8_911c_4842_be08_4674929ed4fb.slice/crio-cc6cddccb4f16a866f544ff5bad1cb8092594e64774ed79855599ba3970bc52f WatchSource:0}: Error finding container cc6cddccb4f16a866f544ff5bad1cb8092594e64774ed79855599ba3970bc52f: Status 404 returned error can't find the container with id cc6cddccb4f16a866f544ff5bad1cb8092594e64774ed79855599ba3970bc52f Oct 04 11:10:51 crc kubenswrapper[4758]: I1004 11:10:51.970690 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 11:10:52 crc kubenswrapper[4758]: I1004 11:10:52.003193 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0e96b041-d262-4169-8ee6-cfee13404759","Type":"ContainerStarted","Data":"5fc1a57cc4c129e283b612d4fbe0ca94bdea937c89097a4e741f0987d47c8cb9"} Oct 04 11:10:52 crc kubenswrapper[4758]: I1004 11:10:52.005675 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"124e2bc8-911c-4842-be08-4674929ed4fb","Type":"ContainerStarted","Data":"cc6cddccb4f16a866f544ff5bad1cb8092594e64774ed79855599ba3970bc52f"} Oct 04 11:10:52 crc kubenswrapper[4758]: I1004 11:10:52.020728 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.02071391 podStartE2EDuration="3.02071391s" podCreationTimestamp="2025-10-04 11:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:52.018647344 +0000 UTC m=+1229.311298233" watchObservedRunningTime="2025-10-04 11:10:52.02071391 +0000 UTC m=+1229.313364799" Oct 04 11:10:53 crc kubenswrapper[4758]: I1004 11:10:53.016530 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"124e2bc8-911c-4842-be08-4674929ed4fb","Type":"ContainerStarted","Data":"2155eee2b141dffefb8d7727ab6829c65e890018300d9e8f4ce95033270b6f50"} Oct 04 11:10:53 crc kubenswrapper[4758]: I1004 11:10:53.016988 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"124e2bc8-911c-4842-be08-4674929ed4fb","Type":"ContainerStarted","Data":"ca15df9b9dd80ad0d22ab7619679a7eaaf97d9dd7ae45f2466027d19a25bfc73"} Oct 04 11:10:53 crc kubenswrapper[4758]: I1004 11:10:53.042628 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.042609299 podStartE2EDuration="2.042609299s" podCreationTimestamp="2025-10-04 11:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:53.036965757 +0000 UTC m=+1230.329616646" watchObservedRunningTime="2025-10-04 11:10:53.042609299 +0000 UTC m=+1230.335260188" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.020290 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.024512 4758 generic.go:334] "Generic (PLEG): container finished" podID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerID="f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877" exitCode=0 Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.025381 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.025541 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a89218c-9de3-466f-9fa5-c6c0cce8053d","Type":"ContainerDied","Data":"f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877"} Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.025570 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a89218c-9de3-466f-9fa5-c6c0cce8053d","Type":"ContainerDied","Data":"6d90478a495aacfb146109a48266a4184aa3d2143d5e1eeb6669895fa87be7dc"} Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.025585 4758 scope.go:117] "RemoveContainer" containerID="f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.056242 4758 scope.go:117] "RemoveContainer" containerID="05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.083714 4758 scope.go:117] "RemoveContainer" containerID="f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.084761 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-combined-ca-bundle\") pod \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.084828 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-config-data\") pod \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.084947 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a89218c-9de3-466f-9fa5-c6c0cce8053d-logs\") pod \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.084975 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85bkj\" (UniqueName: \"kubernetes.io/projected/3a89218c-9de3-466f-9fa5-c6c0cce8053d-kube-api-access-85bkj\") pod \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.085004 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-public-tls-certs\") pod \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.085085 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-internal-tls-certs\") pod \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\" (UID: \"3a89218c-9de3-466f-9fa5-c6c0cce8053d\") " Oct 04 11:10:54 crc kubenswrapper[4758]: E1004 11:10:54.085660 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877\": container with ID starting with f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877 not found: ID does not exist" containerID="f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.085707 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877"} err="failed to get container status \"f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877\": rpc error: code = NotFound desc = could not find container \"f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877\": container with ID starting with f765c46a625bd2d132034a15f419de55e597efad52aa84c001cf2f94eaa65877 not found: ID does not exist" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.085736 4758 scope.go:117] "RemoveContainer" containerID="05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.086855 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a89218c-9de3-466f-9fa5-c6c0cce8053d-logs" (OuterVolumeSpecName: "logs") pod "3a89218c-9de3-466f-9fa5-c6c0cce8053d" (UID: "3a89218c-9de3-466f-9fa5-c6c0cce8053d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.095386 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a89218c-9de3-466f-9fa5-c6c0cce8053d-kube-api-access-85bkj" (OuterVolumeSpecName: "kube-api-access-85bkj") pod "3a89218c-9de3-466f-9fa5-c6c0cce8053d" (UID: "3a89218c-9de3-466f-9fa5-c6c0cce8053d"). InnerVolumeSpecName "kube-api-access-85bkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:54 crc kubenswrapper[4758]: E1004 11:10:54.097851 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0\": container with ID starting with 05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0 not found: ID does not exist" containerID="05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.097914 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0"} err="failed to get container status \"05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0\": rpc error: code = NotFound desc = could not find container \"05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0\": container with ID starting with 05dceca142e469f1404943e32aa35cd468d516983afe8044356164941f97c7b0 not found: ID does not exist" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.150593 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3a89218c-9de3-466f-9fa5-c6c0cce8053d" (UID: "3a89218c-9de3-466f-9fa5-c6c0cce8053d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.153700 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3a89218c-9de3-466f-9fa5-c6c0cce8053d" (UID: "3a89218c-9de3-466f-9fa5-c6c0cce8053d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.161787 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a89218c-9de3-466f-9fa5-c6c0cce8053d" (UID: "3a89218c-9de3-466f-9fa5-c6c0cce8053d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.162988 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-config-data" (OuterVolumeSpecName: "config-data") pod "3a89218c-9de3-466f-9fa5-c6c0cce8053d" (UID: "3a89218c-9de3-466f-9fa5-c6c0cce8053d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.187929 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.188217 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.188302 4758 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a89218c-9de3-466f-9fa5-c6c0cce8053d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.188368 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85bkj\" (UniqueName: \"kubernetes.io/projected/3a89218c-9de3-466f-9fa5-c6c0cce8053d-kube-api-access-85bkj\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.188441 4758 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.188507 4758 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a89218c-9de3-466f-9fa5-c6c0cce8053d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.354032 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.360068 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.378344 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:54 crc kubenswrapper[4758]: E1004 11:10:54.378946 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-api" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.379028 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-api" Oct 04 11:10:54 crc kubenswrapper[4758]: E1004 11:10:54.379135 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-log" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.379201 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-log" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.379421 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-api" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.379488 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" containerName="nova-api-log" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.380462 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.382580 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.382982 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.389784 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.395730 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.492490 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d57de266-f458-4d3b-817c-0c33edfcd1f6-logs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.492719 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.492812 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-config-data\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.492953 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-public-tls-certs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.493043 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdzf6\" (UniqueName: \"kubernetes.io/projected/d57de266-f458-4d3b-817c-0c33edfcd1f6-kube-api-access-vdzf6\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.493126 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.595260 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-public-tls-certs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.595377 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdzf6\" (UniqueName: \"kubernetes.io/projected/d57de266-f458-4d3b-817c-0c33edfcd1f6-kube-api-access-vdzf6\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.595415 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.595489 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d57de266-f458-4d3b-817c-0c33edfcd1f6-logs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.595548 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.595598 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-config-data\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.597973 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d57de266-f458-4d3b-817c-0c33edfcd1f6-logs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.599969 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.599988 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-config-data\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.606293 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.607379 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d57de266-f458-4d3b-817c-0c33edfcd1f6-public-tls-certs\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.620044 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdzf6\" (UniqueName: \"kubernetes.io/projected/d57de266-f458-4d3b-817c-0c33edfcd1f6-kube-api-access-vdzf6\") pod \"nova-api-0\" (UID: \"d57de266-f458-4d3b-817c-0c33edfcd1f6\") " pod="openstack/nova-api-0" Oct 04 11:10:54 crc kubenswrapper[4758]: I1004 11:10:54.700117 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 11:10:55 crc kubenswrapper[4758]: I1004 11:10:55.199811 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 11:10:55 crc kubenswrapper[4758]: W1004 11:10:55.199916 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd57de266_f458_4d3b_817c_0c33edfcd1f6.slice/crio-f035812ffc198038d8cb13517a0732280bfa8e3e5681b6cdbe17889333db60ff WatchSource:0}: Error finding container f035812ffc198038d8cb13517a0732280bfa8e3e5681b6cdbe17889333db60ff: Status 404 returned error can't find the container with id f035812ffc198038d8cb13517a0732280bfa8e3e5681b6cdbe17889333db60ff Oct 04 11:10:55 crc kubenswrapper[4758]: I1004 11:10:55.335357 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a89218c-9de3-466f-9fa5-c6c0cce8053d" path="/var/lib/kubelet/pods/3a89218c-9de3-466f-9fa5-c6c0cce8053d/volumes" Oct 04 11:10:55 crc kubenswrapper[4758]: I1004 11:10:55.360117 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 11:10:56 crc kubenswrapper[4758]: I1004 11:10:56.043682 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d57de266-f458-4d3b-817c-0c33edfcd1f6","Type":"ContainerStarted","Data":"2218578a896a8ce75e8696abb719e426e562c4e8297aa33d4c3090b2174ee585"} Oct 04 11:10:56 crc kubenswrapper[4758]: I1004 11:10:56.043946 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d57de266-f458-4d3b-817c-0c33edfcd1f6","Type":"ContainerStarted","Data":"c315b319e10a110b7402f6c7b4cfe353c300460debcac5d0c2e10dce19ceebe8"} Oct 04 11:10:56 crc kubenswrapper[4758]: I1004 11:10:56.043959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d57de266-f458-4d3b-817c-0c33edfcd1f6","Type":"ContainerStarted","Data":"f035812ffc198038d8cb13517a0732280bfa8e3e5681b6cdbe17889333db60ff"} Oct 04 11:10:56 crc kubenswrapper[4758]: I1004 11:10:56.071159 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.071142177 podStartE2EDuration="2.071142177s" podCreationTimestamp="2025-10-04 11:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:10:56.062535516 +0000 UTC m=+1233.355186415" watchObservedRunningTime="2025-10-04 11:10:56.071142177 +0000 UTC m=+1233.363793066" Oct 04 11:10:56 crc kubenswrapper[4758]: I1004 11:10:56.406972 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 11:10:56 crc kubenswrapper[4758]: I1004 11:10:56.407945 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 11:11:00 crc kubenswrapper[4758]: I1004 11:11:00.360391 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 11:11:00 crc kubenswrapper[4758]: I1004 11:11:00.385500 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.123243 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.249795 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.249862 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.249910 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.250704 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56399f3880bb0482a5458063cd6de09cb352f18aa4d98a39db29016b0272c352"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.250780 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://56399f3880bb0482a5458063cd6de09cb352f18aa4d98a39db29016b0272c352" gracePeriod=600 Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.407455 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 11:11:01 crc kubenswrapper[4758]: I1004 11:11:01.408188 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 11:11:02 crc kubenswrapper[4758]: I1004 11:11:02.103069 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="56399f3880bb0482a5458063cd6de09cb352f18aa4d98a39db29016b0272c352" exitCode=0 Oct 04 11:11:02 crc kubenswrapper[4758]: I1004 11:11:02.103212 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"56399f3880bb0482a5458063cd6de09cb352f18aa4d98a39db29016b0272c352"} Oct 04 11:11:02 crc kubenswrapper[4758]: I1004 11:11:02.103758 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"915ebcfb6b28055c5624cc8d51ba7691c189880fd98368198393745346dfbd64"} Oct 04 11:11:02 crc kubenswrapper[4758]: I1004 11:11:02.103778 4758 scope.go:117] "RemoveContainer" containerID="2b150cc50b08a3a5c33e13466357eafa5f32ab57cc3437792ce3dec5314cb49d" Oct 04 11:11:02 crc kubenswrapper[4758]: I1004 11:11:02.419304 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="124e2bc8-911c-4842-be08-4674929ed4fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:11:02 crc kubenswrapper[4758]: I1004 11:11:02.419335 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="124e2bc8-911c-4842-be08-4674929ed4fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 11:11:04 crc kubenswrapper[4758]: I1004 11:11:04.700852 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:11:04 crc kubenswrapper[4758]: I1004 11:11:04.701527 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 11:11:05 crc kubenswrapper[4758]: I1004 11:11:05.721498 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d57de266-f458-4d3b-817c-0c33edfcd1f6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:11:05 crc kubenswrapper[4758]: I1004 11:11:05.723040 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d57de266-f458-4d3b-817c-0c33edfcd1f6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 11:11:11 crc kubenswrapper[4758]: I1004 11:11:11.423574 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 11:11:11 crc kubenswrapper[4758]: I1004 11:11:11.425838 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 11:11:11 crc kubenswrapper[4758]: I1004 11:11:11.429966 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 11:11:12 crc kubenswrapper[4758]: I1004 11:11:12.218311 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.245225 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.708243 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.708315 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.708933 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.708973 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.719397 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 11:11:14 crc kubenswrapper[4758]: I1004 11:11:14.721764 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 11:11:23 crc kubenswrapper[4758]: I1004 11:11:23.798236 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:11:24 crc kubenswrapper[4758]: I1004 11:11:24.497521 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:11:28 crc kubenswrapper[4758]: I1004 11:11:28.267718 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="rabbitmq" containerID="cri-o://765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e" gracePeriod=604796 Oct 04 11:11:29 crc kubenswrapper[4758]: I1004 11:11:29.021739 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="rabbitmq" containerID="cri-o://77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2" gracePeriod=604796 Oct 04 11:11:30 crc kubenswrapper[4758]: I1004 11:11:30.228410 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 04 11:11:30 crc kubenswrapper[4758]: I1004 11:11:30.508498 4758 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.845943 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955031 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-tls\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955093 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-plugins-conf\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955128 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/acd7d674-14dc-4292-a4b1-5175299225e0-pod-info\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955159 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-erlang-cookie\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955253 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-plugins\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955274 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-confd\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955320 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955339 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-server-conf\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955355 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hbxs\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-kube-api-access-8hbxs\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955410 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-config-data\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.955446 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/acd7d674-14dc-4292-a4b1-5175299225e0-erlang-cookie-secret\") pod \"acd7d674-14dc-4292-a4b1-5175299225e0\" (UID: \"acd7d674-14dc-4292-a4b1-5175299225e0\") " Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.957310 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.957906 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.958246 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.961588 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/acd7d674-14dc-4292-a4b1-5175299225e0-pod-info" (OuterVolumeSpecName: "pod-info") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.962997 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.965318 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd7d674-14dc-4292-a4b1-5175299225e0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.969450 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-kube-api-access-8hbxs" (OuterVolumeSpecName: "kube-api-access-8hbxs") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "kube-api-access-8hbxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.970794 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:11:34 crc kubenswrapper[4758]: I1004 11:11:34.989951 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-config-data" (OuterVolumeSpecName: "config-data") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.046865 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-server-conf" (OuterVolumeSpecName: "server-conf") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057713 4758 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057737 4758 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/acd7d674-14dc-4292-a4b1-5175299225e0-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057746 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057755 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057774 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057782 4758 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057790 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hbxs\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-kube-api-access-8hbxs\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057825 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acd7d674-14dc-4292-a4b1-5175299225e0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057835 4758 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/acd7d674-14dc-4292-a4b1-5175299225e0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.057843 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.083385 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.088555 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "acd7d674-14dc-4292-a4b1-5175299225e0" (UID: "acd7d674-14dc-4292-a4b1-5175299225e0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.159130 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.159165 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/acd7d674-14dc-4292-a4b1-5175299225e0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.363808 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.441128 4758 generic.go:334] "Generic (PLEG): container finished" podID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerID="77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2" exitCode=0 Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.441219 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"57d6ac32-db38-4443-9c0c-13c66db2451b","Type":"ContainerDied","Data":"77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2"} Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.441249 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"57d6ac32-db38-4443-9c0c-13c66db2451b","Type":"ContainerDied","Data":"ecb705bc249ee000fc5a802b33bccf037c245032292d7fac85d80c388236539e"} Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.441283 4758 scope.go:117] "RemoveContainer" containerID="77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.441452 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.443553 4758 generic.go:334] "Generic (PLEG): container finished" podID="acd7d674-14dc-4292-a4b1-5175299225e0" containerID="765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e" exitCode=0 Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.443579 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"acd7d674-14dc-4292-a4b1-5175299225e0","Type":"ContainerDied","Data":"765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e"} Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.443596 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"acd7d674-14dc-4292-a4b1-5175299225e0","Type":"ContainerDied","Data":"354b9d7f44922144c326175be25e0116e0de5eee93bfe37729b5ad5ee9f3aba6"} Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.443664 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467028 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467088 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57d6ac32-db38-4443-9c0c-13c66db2451b-erlang-cookie-secret\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467208 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-plugins-conf\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467243 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-plugins\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467264 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-config-data\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467287 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-confd\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467325 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57d6ac32-db38-4443-9c0c-13c66db2451b-pod-info\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467351 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-server-conf\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467375 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-tls\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467401 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-erlang-cookie\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.467445 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5m7kn\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-kube-api-access-5m7kn\") pod \"57d6ac32-db38-4443-9c0c-13c66db2451b\" (UID: \"57d6ac32-db38-4443-9c0c-13c66db2451b\") " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.475895 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.476040 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.476269 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.477083 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.480201 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.481933 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-kube-api-access-5m7kn" (OuterVolumeSpecName: "kube-api-access-5m7kn") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "kube-api-access-5m7kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.482910 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.483208 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57d6ac32-db38-4443-9c0c-13c66db2451b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.498435 4758 scope.go:117] "RemoveContainer" containerID="a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.499188 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/57d6ac32-db38-4443-9c0c-13c66db2451b-pod-info" (OuterVolumeSpecName: "pod-info") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.524861 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.545271 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.546270 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="rabbitmq" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.546294 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="rabbitmq" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.546307 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="setup-container" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.546319 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="setup-container" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.546334 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="rabbitmq" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.546340 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="rabbitmq" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.546364 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="setup-container" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.546390 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="setup-container" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.546597 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" containerName="rabbitmq" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.546616 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" containerName="rabbitmq" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.549522 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.551286 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.552716 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.553695 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.553749 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.553872 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.553941 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ctr8l" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.554017 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.560131 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.562981 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-config-data" (OuterVolumeSpecName: "config-data") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.596009 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.596754 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.596890 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.596920 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.596964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec7b1391-1a82-4e8b-ba1d-131842a3908c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.596986 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597026 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597061 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597111 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec7b1391-1a82-4e8b-ba1d-131842a3908c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597153 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqvqx\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-kube-api-access-vqvqx\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597289 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597664 4758 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.597963 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598384 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598395 4758 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/57d6ac32-db38-4443-9c0c-13c66db2451b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598402 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598412 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598527 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5m7kn\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-kube-api-access-5m7kn\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598547 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.598557 4758 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/57d6ac32-db38-4443-9c0c-13c66db2451b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.670662 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.672228 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.676166 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-server-conf" (OuterVolumeSpecName: "server-conf") pod "57d6ac32-db38-4443-9c0c-13c66db2451b" (UID: "57d6ac32-db38-4443-9c0c-13c66db2451b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.695748 4758 scope.go:117] "RemoveContainer" containerID="77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.697054 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2\": container with ID starting with 77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2 not found: ID does not exist" containerID="77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.697169 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2"} err="failed to get container status \"77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2\": rpc error: code = NotFound desc = could not find container \"77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2\": container with ID starting with 77e8cbbcc752bb49b384456a7b30c20d6c27d4b96663c1cda7394c021f1039f2 not found: ID does not exist" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.697208 4758 scope.go:117] "RemoveContainer" containerID="a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.697602 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17\": container with ID starting with a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17 not found: ID does not exist" containerID="a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.697649 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17"} err="failed to get container status \"a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17\": rpc error: code = NotFound desc = could not find container \"a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17\": container with ID starting with a9f55b3aae61fef26bec5ee0a239a5459bc7aea03cadf627a6cb5ee4b4fe4d17 not found: ID does not exist" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.697675 4758 scope.go:117] "RemoveContainer" containerID="765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700281 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqvqx\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-kube-api-access-vqvqx\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700326 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700396 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700422 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700487 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700532 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700572 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec7b1391-1a82-4e8b-ba1d-131842a3908c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700613 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700630 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700649 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700691 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec7b1391-1a82-4e8b-ba1d-131842a3908c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700764 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700776 4758 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/57d6ac32-db38-4443-9c0c-13c66db2451b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.700785 4758 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/57d6ac32-db38-4443-9c0c-13c66db2451b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.702985 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.703227 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.704599 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ec7b1391-1a82-4e8b-ba1d-131842a3908c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.704742 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.706316 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.706594 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.707298 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec7b1391-1a82-4e8b-ba1d-131842a3908c-config-data\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.711484 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.713040 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.720692 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ec7b1391-1a82-4e8b-ba1d-131842a3908c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.724747 4758 scope.go:117] "RemoveContainer" containerID="fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.726062 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqvqx\" (UniqueName: \"kubernetes.io/projected/ec7b1391-1a82-4e8b-ba1d-131842a3908c-kube-api-access-vqvqx\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.749603 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"ec7b1391-1a82-4e8b-ba1d-131842a3908c\") " pod="openstack/rabbitmq-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.756698 4758 scope.go:117] "RemoveContainer" containerID="765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.759050 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e\": container with ID starting with 765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e not found: ID does not exist" containerID="765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.759085 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e"} err="failed to get container status \"765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e\": rpc error: code = NotFound desc = could not find container \"765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e\": container with ID starting with 765b63fa461c442d66e76ef690108dd5a6772ceb0457347ed051e033de0dac1e not found: ID does not exist" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.759124 4758 scope.go:117] "RemoveContainer" containerID="fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4" Oct 04 11:11:35 crc kubenswrapper[4758]: E1004 11:11:35.759535 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4\": container with ID starting with fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4 not found: ID does not exist" containerID="fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.759591 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4"} err="failed to get container status \"fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4\": rpc error: code = NotFound desc = could not find container \"fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4\": container with ID starting with fb9c9a3f7bd6d74dbe3def29fe4dc789ed6dede5e8e75af36f9fc620798d99e4 not found: ID does not exist" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.785299 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.793179 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.807102 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.808647 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.810856 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.815254 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.815284 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gvv74" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.815399 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.819061 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.820809 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.821055 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.832375 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.904720 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg55l\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-kube-api-access-pg55l\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.905884 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906071 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906194 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906303 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906425 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906536 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906653 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906741 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906850 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.906971 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:35 crc kubenswrapper[4758]: I1004 11:11:35.997429 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008725 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008781 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008807 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008826 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008846 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008880 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008934 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg55l\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-kube-api-access-pg55l\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.008996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.009013 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.009031 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.010045 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.010350 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.011404 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.011738 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.012199 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.012818 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.014534 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.016351 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.018542 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.026135 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.029080 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg55l\" (UniqueName: \"kubernetes.io/projected/dfdc4c39-424b-4488-9f1a-1efe8f212ac2-kube-api-access-pg55l\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.045965 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"dfdc4c39-424b-4488-9f1a-1efe8f212ac2\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.127867 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.268937 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-5qqt5"] Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.296726 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.310052 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.347777 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.347829 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.347858 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv452\" (UniqueName: \"kubernetes.io/projected/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-kube-api-access-zv452\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.347895 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-config\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.347928 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.347971 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.348058 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.365344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-5qqt5"] Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.454471 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-config\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.454702 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.454732 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.454799 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.455620 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.456088 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-config\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.456767 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.457653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.457686 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.457734 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv452\" (UniqueName: \"kubernetes.io/projected/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-kube-api-access-zv452\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.458428 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.458918 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.460937 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.469464 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.480499 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv452\" (UniqueName: \"kubernetes.io/projected/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-kube-api-access-zv452\") pod \"dnsmasq-dns-79bd4cc8c9-5qqt5\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.676303 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 11:11:36 crc kubenswrapper[4758]: I1004 11:11:36.678472 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.176677 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-5qqt5"] Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.335264 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57d6ac32-db38-4443-9c0c-13c66db2451b" path="/var/lib/kubelet/pods/57d6ac32-db38-4443-9c0c-13c66db2451b/volumes" Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.336302 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd7d674-14dc-4292-a4b1-5175299225e0" path="/var/lib/kubelet/pods/acd7d674-14dc-4292-a4b1-5175299225e0/volumes" Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.480250 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec7b1391-1a82-4e8b-ba1d-131842a3908c","Type":"ContainerStarted","Data":"601dfc81ed983876a04065004f7c8814a98115e8f08166c2ee3d61536993a2a0"} Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.480622 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec7b1391-1a82-4e8b-ba1d-131842a3908c","Type":"ContainerStarted","Data":"6ef86370d5c49984190a2a166f152a3af200ed542566abc46de670e59ea5b5d6"} Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.481984 4758 generic.go:334] "Generic (PLEG): container finished" podID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerID="4a1f6d9da24f823ba901b0414efabe972e8f0d93c88ed6f36f52b6f7ab4c0a60" exitCode=0 Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.482039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" event={"ID":"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17","Type":"ContainerDied","Data":"4a1f6d9da24f823ba901b0414efabe972e8f0d93c88ed6f36f52b6f7ab4c0a60"} Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.482060 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" event={"ID":"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17","Type":"ContainerStarted","Data":"0324a1020a1826a5d9b5e9ce898fb7d9acf4a1a17daa88d098e0f2bdb463a6d4"} Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.484496 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dfdc4c39-424b-4488-9f1a-1efe8f212ac2","Type":"ContainerStarted","Data":"d40e367c3426f66ef7c71870fbc62b5dd7b4900c7ca9c5eb6acb33ceac079372"} Oct 04 11:11:37 crc kubenswrapper[4758]: I1004 11:11:37.484534 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dfdc4c39-424b-4488-9f1a-1efe8f212ac2","Type":"ContainerStarted","Data":"cf4070f867766000ec426e0b0e88e22f2d39c1cd65f3c68fa9e1ae652800ecb2"} Oct 04 11:11:38 crc kubenswrapper[4758]: I1004 11:11:38.503807 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" event={"ID":"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17","Type":"ContainerStarted","Data":"4ed00588687bfe0dc3af274d7e501d4dcb8a980e3c08aa8c8c409fc67db38fba"} Oct 04 11:11:38 crc kubenswrapper[4758]: I1004 11:11:38.531513 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" podStartSLOduration=2.531487074 podStartE2EDuration="2.531487074s" podCreationTimestamp="2025-10-04 11:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:11:38.526987804 +0000 UTC m=+1275.819638693" watchObservedRunningTime="2025-10-04 11:11:38.531487074 +0000 UTC m=+1275.824137993" Oct 04 11:11:39 crc kubenswrapper[4758]: I1004 11:11:39.515097 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:46 crc kubenswrapper[4758]: I1004 11:11:46.681533 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:46 crc kubenswrapper[4758]: I1004 11:11:46.756497 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-xbhdg"] Oct 04 11:11:46 crc kubenswrapper[4758]: I1004 11:11:46.756799 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerName="dnsmasq-dns" containerID="cri-o://cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658" gracePeriod=10 Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.081461 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54ffdb7d8c-fz8h2"] Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.083130 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.099964 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.100030 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-ovsdbserver-nb\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.100049 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-dns-swift-storage-0\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.100067 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-dns-svc\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.100096 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-ovsdbserver-sb\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.100225 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggk9m\" (UniqueName: \"kubernetes.io/projected/0acd8097-4dd5-4456-b083-97412d714eb6-kube-api-access-ggk9m\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.100242 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-config\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.108068 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54ffdb7d8c-fz8h2"] Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.201816 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggk9m\" (UniqueName: \"kubernetes.io/projected/0acd8097-4dd5-4456-b083-97412d714eb6-kube-api-access-ggk9m\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.201857 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-config\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.201942 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.201980 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-ovsdbserver-nb\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.202006 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-dns-swift-storage-0\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.202025 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-dns-svc\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.202053 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-ovsdbserver-sb\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.203021 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-ovsdbserver-sb\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.203599 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-openstack-edpm-ipam\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.203675 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-dns-swift-storage-0\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.204255 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-ovsdbserver-nb\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.204490 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-config\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.204784 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0acd8097-4dd5-4456-b083-97412d714eb6-dns-svc\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.237974 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggk9m\" (UniqueName: \"kubernetes.io/projected/0acd8097-4dd5-4456-b083-97412d714eb6-kube-api-access-ggk9m\") pod \"dnsmasq-dns-54ffdb7d8c-fz8h2\" (UID: \"0acd8097-4dd5-4456-b083-97412d714eb6\") " pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.300906 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.405843 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-swift-storage-0\") pod \"ad4df944-5c12-46c7-8791-fc7aa5085e23\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.405924 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-config\") pod \"ad4df944-5c12-46c7-8791-fc7aa5085e23\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.405948 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-sb\") pod \"ad4df944-5c12-46c7-8791-fc7aa5085e23\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.406016 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-nb\") pod \"ad4df944-5c12-46c7-8791-fc7aa5085e23\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.406038 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ntl4\" (UniqueName: \"kubernetes.io/projected/ad4df944-5c12-46c7-8791-fc7aa5085e23-kube-api-access-9ntl4\") pod \"ad4df944-5c12-46c7-8791-fc7aa5085e23\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.406092 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-svc\") pod \"ad4df944-5c12-46c7-8791-fc7aa5085e23\" (UID: \"ad4df944-5c12-46c7-8791-fc7aa5085e23\") " Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.411053 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad4df944-5c12-46c7-8791-fc7aa5085e23-kube-api-access-9ntl4" (OuterVolumeSpecName: "kube-api-access-9ntl4") pod "ad4df944-5c12-46c7-8791-fc7aa5085e23" (UID: "ad4df944-5c12-46c7-8791-fc7aa5085e23"). InnerVolumeSpecName "kube-api-access-9ntl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.411355 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.462845 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-config" (OuterVolumeSpecName: "config") pod "ad4df944-5c12-46c7-8791-fc7aa5085e23" (UID: "ad4df944-5c12-46c7-8791-fc7aa5085e23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.463159 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ad4df944-5c12-46c7-8791-fc7aa5085e23" (UID: "ad4df944-5c12-46c7-8791-fc7aa5085e23"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.476709 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ad4df944-5c12-46c7-8791-fc7aa5085e23" (UID: "ad4df944-5c12-46c7-8791-fc7aa5085e23"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.482556 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ad4df944-5c12-46c7-8791-fc7aa5085e23" (UID: "ad4df944-5c12-46c7-8791-fc7aa5085e23"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.485167 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ad4df944-5c12-46c7-8791-fc7aa5085e23" (UID: "ad4df944-5c12-46c7-8791-fc7aa5085e23"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.508799 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.508828 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.508960 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.508990 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.508998 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad4df944-5c12-46c7-8791-fc7aa5085e23-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.509006 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ntl4\" (UniqueName: \"kubernetes.io/projected/ad4df944-5c12-46c7-8791-fc7aa5085e23-kube-api-access-9ntl4\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.597581 4758 generic.go:334] "Generic (PLEG): container finished" podID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerID="cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658" exitCode=0 Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.597621 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" event={"ID":"ad4df944-5c12-46c7-8791-fc7aa5085e23","Type":"ContainerDied","Data":"cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658"} Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.597647 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" event={"ID":"ad4df944-5c12-46c7-8791-fc7aa5085e23","Type":"ContainerDied","Data":"fb23bd6268e80256568597d79594330de37c4262f00f3750da3209fe61e62d8f"} Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.597662 4758 scope.go:117] "RemoveContainer" containerID="cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.597776 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-xbhdg" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.635179 4758 scope.go:117] "RemoveContainer" containerID="007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.642224 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-xbhdg"] Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.651083 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-xbhdg"] Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.661782 4758 scope.go:117] "RemoveContainer" containerID="cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658" Oct 04 11:11:47 crc kubenswrapper[4758]: E1004 11:11:47.662346 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658\": container with ID starting with cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658 not found: ID does not exist" containerID="cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.662378 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658"} err="failed to get container status \"cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658\": rpc error: code = NotFound desc = could not find container \"cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658\": container with ID starting with cc05ff6596f4518acb3330eb1a653b5fe6c9b5d732b072b571021fde42ec6658 not found: ID does not exist" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.662397 4758 scope.go:117] "RemoveContainer" containerID="007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53" Oct 04 11:11:47 crc kubenswrapper[4758]: E1004 11:11:47.662612 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53\": container with ID starting with 007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53 not found: ID does not exist" containerID="007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53" Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.662634 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53"} err="failed to get container status \"007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53\": rpc error: code = NotFound desc = could not find container \"007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53\": container with ID starting with 007367f23d97d9270913c5bfe9c5ee729e5a5000879aa67cd1140009f2cffc53 not found: ID does not exist" Oct 04 11:11:47 crc kubenswrapper[4758]: W1004 11:11:47.849130 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0acd8097_4dd5_4456_b083_97412d714eb6.slice/crio-6bc56009854205c719ab80ea03ec4169050ac412d3c3c36280d07f349043716c WatchSource:0}: Error finding container 6bc56009854205c719ab80ea03ec4169050ac412d3c3c36280d07f349043716c: Status 404 returned error can't find the container with id 6bc56009854205c719ab80ea03ec4169050ac412d3c3c36280d07f349043716c Oct 04 11:11:47 crc kubenswrapper[4758]: I1004 11:11:47.849803 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54ffdb7d8c-fz8h2"] Oct 04 11:11:48 crc kubenswrapper[4758]: I1004 11:11:48.609417 4758 generic.go:334] "Generic (PLEG): container finished" podID="0acd8097-4dd5-4456-b083-97412d714eb6" containerID="4e099579569f2b0e68ce8709165776ddb640bc8c50f0292c337614ff04f782e4" exitCode=0 Oct 04 11:11:48 crc kubenswrapper[4758]: I1004 11:11:48.609496 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" event={"ID":"0acd8097-4dd5-4456-b083-97412d714eb6","Type":"ContainerDied","Data":"4e099579569f2b0e68ce8709165776ddb640bc8c50f0292c337614ff04f782e4"} Oct 04 11:11:48 crc kubenswrapper[4758]: I1004 11:11:48.609698 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" event={"ID":"0acd8097-4dd5-4456-b083-97412d714eb6","Type":"ContainerStarted","Data":"6bc56009854205c719ab80ea03ec4169050ac412d3c3c36280d07f349043716c"} Oct 04 11:11:49 crc kubenswrapper[4758]: I1004 11:11:49.337685 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" path="/var/lib/kubelet/pods/ad4df944-5c12-46c7-8791-fc7aa5085e23/volumes" Oct 04 11:11:49 crc kubenswrapper[4758]: I1004 11:11:49.622281 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" event={"ID":"0acd8097-4dd5-4456-b083-97412d714eb6","Type":"ContainerStarted","Data":"ed1f5b803d2b6dfaeda1929549cbc9e46cc9e306539ce9f27545fc8573fd8476"} Oct 04 11:11:49 crc kubenswrapper[4758]: I1004 11:11:49.622674 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:49 crc kubenswrapper[4758]: I1004 11:11:49.641722 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" podStartSLOduration=2.64170008 podStartE2EDuration="2.64170008s" podCreationTimestamp="2025-10-04 11:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:11:49.640374485 +0000 UTC m=+1286.933025394" watchObservedRunningTime="2025-10-04 11:11:49.64170008 +0000 UTC m=+1286.934350979" Oct 04 11:11:57 crc kubenswrapper[4758]: I1004 11:11:57.412371 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54ffdb7d8c-fz8h2" Oct 04 11:11:57 crc kubenswrapper[4758]: I1004 11:11:57.503373 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-5qqt5"] Oct 04 11:11:57 crc kubenswrapper[4758]: I1004 11:11:57.503686 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerName="dnsmasq-dns" containerID="cri-o://4ed00588687bfe0dc3af274d7e501d4dcb8a980e3c08aa8c8c409fc67db38fba" gracePeriod=10 Oct 04 11:11:57 crc kubenswrapper[4758]: I1004 11:11:57.693996 4758 generic.go:334] "Generic (PLEG): container finished" podID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerID="4ed00588687bfe0dc3af274d7e501d4dcb8a980e3c08aa8c8c409fc67db38fba" exitCode=0 Oct 04 11:11:57 crc kubenswrapper[4758]: I1004 11:11:57.694035 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" event={"ID":"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17","Type":"ContainerDied","Data":"4ed00588687bfe0dc3af274d7e501d4dcb8a980e3c08aa8c8c409fc67db38fba"} Oct 04 11:11:57 crc kubenswrapper[4758]: I1004 11:11:57.986978 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.009332 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-svc\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.009381 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-sb\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.009452 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-config\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.092290 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.097629 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-config" (OuterVolumeSpecName: "config") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.104122 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.110607 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-openstack-edpm-ipam\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.110669 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-nb\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.110870 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-swift-storage-0\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.110917 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv452\" (UniqueName: \"kubernetes.io/projected/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-kube-api-access-zv452\") pod \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\" (UID: \"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17\") " Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.111309 4758 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.111326 4758 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.111335 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.117465 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-kube-api-access-zv452" (OuterVolumeSpecName: "kube-api-access-zv452") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "kube-api-access-zv452". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.159789 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.161256 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.165218 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" (UID: "16f3bcf3-1ab6-45e3-8759-93b1ca52dc17"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.212734 4758 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.212765 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv452\" (UniqueName: \"kubernetes.io/projected/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-kube-api-access-zv452\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.212774 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.212782 4758 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.705914 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" event={"ID":"16f3bcf3-1ab6-45e3-8759-93b1ca52dc17","Type":"ContainerDied","Data":"0324a1020a1826a5d9b5e9ce898fb7d9acf4a1a17daa88d098e0f2bdb463a6d4"} Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.705968 4758 scope.go:117] "RemoveContainer" containerID="4ed00588687bfe0dc3af274d7e501d4dcb8a980e3c08aa8c8c409fc67db38fba" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.705989 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-5qqt5" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.744863 4758 scope.go:117] "RemoveContainer" containerID="4a1f6d9da24f823ba901b0414efabe972e8f0d93c88ed6f36f52b6f7ab4c0a60" Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.747930 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-5qqt5"] Oct 04 11:11:58 crc kubenswrapper[4758]: I1004 11:11:58.754611 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-5qqt5"] Oct 04 11:11:59 crc kubenswrapper[4758]: I1004 11:11:59.344000 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" path="/var/lib/kubelet/pods/16f3bcf3-1ab6-45e3-8759-93b1ca52dc17/volumes" Oct 04 11:12:07 crc kubenswrapper[4758]: I1004 11:12:07.800547 4758 generic.go:334] "Generic (PLEG): container finished" podID="dfdc4c39-424b-4488-9f1a-1efe8f212ac2" containerID="d40e367c3426f66ef7c71870fbc62b5dd7b4900c7ca9c5eb6acb33ceac079372" exitCode=0 Oct 04 11:12:07 crc kubenswrapper[4758]: I1004 11:12:07.800665 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dfdc4c39-424b-4488-9f1a-1efe8f212ac2","Type":"ContainerDied","Data":"d40e367c3426f66ef7c71870fbc62b5dd7b4900c7ca9c5eb6acb33ceac079372"} Oct 04 11:12:07 crc kubenswrapper[4758]: I1004 11:12:07.805351 4758 generic.go:334] "Generic (PLEG): container finished" podID="ec7b1391-1a82-4e8b-ba1d-131842a3908c" containerID="601dfc81ed983876a04065004f7c8814a98115e8f08166c2ee3d61536993a2a0" exitCode=0 Oct 04 11:12:07 crc kubenswrapper[4758]: I1004 11:12:07.805395 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec7b1391-1a82-4e8b-ba1d-131842a3908c","Type":"ContainerDied","Data":"601dfc81ed983876a04065004f7c8814a98115e8f08166c2ee3d61536993a2a0"} Oct 04 11:12:08 crc kubenswrapper[4758]: I1004 11:12:08.815449 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ec7b1391-1a82-4e8b-ba1d-131842a3908c","Type":"ContainerStarted","Data":"3a074246d35f4ea60622a8bc58aa3948e2be205f1552a2be8c461df8b2aeca30"} Oct 04 11:12:08 crc kubenswrapper[4758]: I1004 11:12:08.816338 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 11:12:08 crc kubenswrapper[4758]: I1004 11:12:08.818302 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dfdc4c39-424b-4488-9f1a-1efe8f212ac2","Type":"ContainerStarted","Data":"992ef8ddc1e021b49d841ad30ef5194476e3b4f9ac59cc7d67402b6160509ff0"} Oct 04 11:12:08 crc kubenswrapper[4758]: I1004 11:12:08.818500 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:12:08 crc kubenswrapper[4758]: I1004 11:12:08.849904 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=33.849888307 podStartE2EDuration="33.849888307s" podCreationTimestamp="2025-10-04 11:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:12:08.842470518 +0000 UTC m=+1306.135121407" watchObservedRunningTime="2025-10-04 11:12:08.849888307 +0000 UTC m=+1306.142539196" Oct 04 11:12:08 crc kubenswrapper[4758]: I1004 11:12:08.876880 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=33.876863818 podStartE2EDuration="33.876863818s" podCreationTimestamp="2025-10-04 11:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:12:08.871841914 +0000 UTC m=+1306.164492803" watchObservedRunningTime="2025-10-04 11:12:08.876863818 +0000 UTC m=+1306.169514707" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.855227 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n"] Oct 04 11:12:20 crc kubenswrapper[4758]: E1004 11:12:20.856361 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerName="dnsmasq-dns" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.856382 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerName="dnsmasq-dns" Oct 04 11:12:20 crc kubenswrapper[4758]: E1004 11:12:20.856424 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerName="dnsmasq-dns" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.856438 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerName="dnsmasq-dns" Oct 04 11:12:20 crc kubenswrapper[4758]: E1004 11:12:20.856468 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerName="init" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.856481 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerName="init" Oct 04 11:12:20 crc kubenswrapper[4758]: E1004 11:12:20.856521 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerName="init" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.856534 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerName="init" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.856852 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad4df944-5c12-46c7-8791-fc7aa5085e23" containerName="dnsmasq-dns" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.856879 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f3bcf3-1ab6-45e3-8759-93b1ca52dc17" containerName="dnsmasq-dns" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.857878 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.864334 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.864704 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.865165 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.869008 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.872977 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n"] Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.935048 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9ml\" (UniqueName: \"kubernetes.io/projected/daefc3f9-3b67-4223-844d-025717e3094d-kube-api-access-xx9ml\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.935185 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.935333 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:20 crc kubenswrapper[4758]: I1004 11:12:20.935382 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.036590 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.036653 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.036719 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9ml\" (UniqueName: \"kubernetes.io/projected/daefc3f9-3b67-4223-844d-025717e3094d-kube-api-access-xx9ml\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.036756 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.060812 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.061175 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.064701 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.077737 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9ml\" (UniqueName: \"kubernetes.io/projected/daefc3f9-3b67-4223-844d-025717e3094d-kube-api-access-xx9ml\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.180344 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:21 crc kubenswrapper[4758]: I1004 11:12:21.942297 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n"] Oct 04 11:12:21 crc kubenswrapper[4758]: W1004 11:12:21.949634 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaefc3f9_3b67_4223_844d_025717e3094d.slice/crio-51023cd3a2bdbabcfaadb0d0f1232a0c7246a2dba0b7794f5fac5df997a59914 WatchSource:0}: Error finding container 51023cd3a2bdbabcfaadb0d0f1232a0c7246a2dba0b7794f5fac5df997a59914: Status 404 returned error can't find the container with id 51023cd3a2bdbabcfaadb0d0f1232a0c7246a2dba0b7794f5fac5df997a59914 Oct 04 11:12:22 crc kubenswrapper[4758]: I1004 11:12:22.959732 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" event={"ID":"daefc3f9-3b67-4223-844d-025717e3094d","Type":"ContainerStarted","Data":"51023cd3a2bdbabcfaadb0d0f1232a0c7246a2dba0b7794f5fac5df997a59914"} Oct 04 11:12:26 crc kubenswrapper[4758]: I1004 11:12:26.008177 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 11:12:26 crc kubenswrapper[4758]: I1004 11:12:26.149568 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 11:12:32 crc kubenswrapper[4758]: I1004 11:12:32.862624 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:12:34 crc kubenswrapper[4758]: I1004 11:12:34.073259 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" event={"ID":"daefc3f9-3b67-4223-844d-025717e3094d","Type":"ContainerStarted","Data":"66ae3be66b25790a7d275ad656573e5ce152c56da58a60b3e6ee8f5953e0f425"} Oct 04 11:12:34 crc kubenswrapper[4758]: I1004 11:12:34.141675 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" podStartSLOduration=3.233541163 podStartE2EDuration="14.141653943s" podCreationTimestamp="2025-10-04 11:12:20 +0000 UTC" firstStartedPulling="2025-10-04 11:12:21.951867279 +0000 UTC m=+1319.244518178" lastFinishedPulling="2025-10-04 11:12:32.859980069 +0000 UTC m=+1330.152630958" observedRunningTime="2025-10-04 11:12:34.130746321 +0000 UTC m=+1331.423397210" watchObservedRunningTime="2025-10-04 11:12:34.141653943 +0000 UTC m=+1331.434304832" Oct 04 11:12:44 crc kubenswrapper[4758]: I1004 11:12:44.178729 4758 generic.go:334] "Generic (PLEG): container finished" podID="daefc3f9-3b67-4223-844d-025717e3094d" containerID="66ae3be66b25790a7d275ad656573e5ce152c56da58a60b3e6ee8f5953e0f425" exitCode=0 Oct 04 11:12:44 crc kubenswrapper[4758]: I1004 11:12:44.179216 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" event={"ID":"daefc3f9-3b67-4223-844d-025717e3094d","Type":"ContainerDied","Data":"66ae3be66b25790a7d275ad656573e5ce152c56da58a60b3e6ee8f5953e0f425"} Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.726652 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.864984 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-repo-setup-combined-ca-bundle\") pod \"daefc3f9-3b67-4223-844d-025717e3094d\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.865030 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx9ml\" (UniqueName: \"kubernetes.io/projected/daefc3f9-3b67-4223-844d-025717e3094d-kube-api-access-xx9ml\") pod \"daefc3f9-3b67-4223-844d-025717e3094d\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.865238 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-inventory\") pod \"daefc3f9-3b67-4223-844d-025717e3094d\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.865269 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-ssh-key\") pod \"daefc3f9-3b67-4223-844d-025717e3094d\" (UID: \"daefc3f9-3b67-4223-844d-025717e3094d\") " Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.880289 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daefc3f9-3b67-4223-844d-025717e3094d-kube-api-access-xx9ml" (OuterVolumeSpecName: "kube-api-access-xx9ml") pod "daefc3f9-3b67-4223-844d-025717e3094d" (UID: "daefc3f9-3b67-4223-844d-025717e3094d"). InnerVolumeSpecName "kube-api-access-xx9ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.881321 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "daefc3f9-3b67-4223-844d-025717e3094d" (UID: "daefc3f9-3b67-4223-844d-025717e3094d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.890810 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-inventory" (OuterVolumeSpecName: "inventory") pod "daefc3f9-3b67-4223-844d-025717e3094d" (UID: "daefc3f9-3b67-4223-844d-025717e3094d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.893788 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "daefc3f9-3b67-4223-844d-025717e3094d" (UID: "daefc3f9-3b67-4223-844d-025717e3094d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.967825 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.968025 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.968134 4758 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daefc3f9-3b67-4223-844d-025717e3094d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:45 crc kubenswrapper[4758]: I1004 11:12:45.968248 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx9ml\" (UniqueName: \"kubernetes.io/projected/daefc3f9-3b67-4223-844d-025717e3094d-kube-api-access-xx9ml\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.200029 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" event={"ID":"daefc3f9-3b67-4223-844d-025717e3094d","Type":"ContainerDied","Data":"51023cd3a2bdbabcfaadb0d0f1232a0c7246a2dba0b7794f5fac5df997a59914"} Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.200072 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51023cd3a2bdbabcfaadb0d0f1232a0c7246a2dba0b7794f5fac5df997a59914" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.200155 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.355800 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t"] Oct 04 11:12:46 crc kubenswrapper[4758]: E1004 11:12:46.356253 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daefc3f9-3b67-4223-844d-025717e3094d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.356270 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="daefc3f9-3b67-4223-844d-025717e3094d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.356461 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="daefc3f9-3b67-4223-844d-025717e3094d" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.357074 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.358975 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.361940 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.362138 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.362273 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.374581 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t"] Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.479257 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.479793 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgdzb\" (UniqueName: \"kubernetes.io/projected/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-kube-api-access-fgdzb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.479954 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.581291 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.581543 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgdzb\" (UniqueName: \"kubernetes.io/projected/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-kube-api-access-fgdzb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.581670 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.593715 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.594856 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.634084 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgdzb\" (UniqueName: \"kubernetes.io/projected/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-kube-api-access-fgdzb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-zml5t\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:46 crc kubenswrapper[4758]: I1004 11:12:46.673715 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:47 crc kubenswrapper[4758]: I1004 11:12:47.239474 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t"] Oct 04 11:12:48 crc kubenswrapper[4758]: I1004 11:12:48.223873 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" event={"ID":"b7817b9e-b878-40b6-b1a5-9ed2526c32a5","Type":"ContainerStarted","Data":"0af8edcb71d58126f203c511499a4b26b6f0db81d4b922b730621b257bbda5e6"} Oct 04 11:12:48 crc kubenswrapper[4758]: I1004 11:12:48.224444 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" event={"ID":"b7817b9e-b878-40b6-b1a5-9ed2526c32a5","Type":"ContainerStarted","Data":"3bca4b7635266372118c4ec1cde794aa72ace2bcb78fafe7f987ebf96ff4b350"} Oct 04 11:12:48 crc kubenswrapper[4758]: I1004 11:12:48.252613 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" podStartSLOduration=1.8068602710000001 podStartE2EDuration="2.252588857s" podCreationTimestamp="2025-10-04 11:12:46 +0000 UTC" firstStartedPulling="2025-10-04 11:12:47.243702496 +0000 UTC m=+1344.536353395" lastFinishedPulling="2025-10-04 11:12:47.689431042 +0000 UTC m=+1344.982081981" observedRunningTime="2025-10-04 11:12:48.237386821 +0000 UTC m=+1345.530037710" watchObservedRunningTime="2025-10-04 11:12:48.252588857 +0000 UTC m=+1345.545239786" Oct 04 11:12:51 crc kubenswrapper[4758]: I1004 11:12:51.262705 4758 generic.go:334] "Generic (PLEG): container finished" podID="b7817b9e-b878-40b6-b1a5-9ed2526c32a5" containerID="0af8edcb71d58126f203c511499a4b26b6f0db81d4b922b730621b257bbda5e6" exitCode=0 Oct 04 11:12:51 crc kubenswrapper[4758]: I1004 11:12:51.262838 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" event={"ID":"b7817b9e-b878-40b6-b1a5-9ed2526c32a5","Type":"ContainerDied","Data":"0af8edcb71d58126f203c511499a4b26b6f0db81d4b922b730621b257bbda5e6"} Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.689021 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.825668 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-inventory\") pod \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.825803 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgdzb\" (UniqueName: \"kubernetes.io/projected/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-kube-api-access-fgdzb\") pod \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.825852 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-ssh-key\") pod \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\" (UID: \"b7817b9e-b878-40b6-b1a5-9ed2526c32a5\") " Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.830558 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-kube-api-access-fgdzb" (OuterVolumeSpecName: "kube-api-access-fgdzb") pod "b7817b9e-b878-40b6-b1a5-9ed2526c32a5" (UID: "b7817b9e-b878-40b6-b1a5-9ed2526c32a5"). InnerVolumeSpecName "kube-api-access-fgdzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.850636 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b7817b9e-b878-40b6-b1a5-9ed2526c32a5" (UID: "b7817b9e-b878-40b6-b1a5-9ed2526c32a5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.851272 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-inventory" (OuterVolumeSpecName: "inventory") pod "b7817b9e-b878-40b6-b1a5-9ed2526c32a5" (UID: "b7817b9e-b878-40b6-b1a5-9ed2526c32a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.928086 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.928126 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgdzb\" (UniqueName: \"kubernetes.io/projected/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-kube-api-access-fgdzb\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:52 crc kubenswrapper[4758]: I1004 11:12:52.928137 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b7817b9e-b878-40b6-b1a5-9ed2526c32a5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.280255 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" event={"ID":"b7817b9e-b878-40b6-b1a5-9ed2526c32a5","Type":"ContainerDied","Data":"3bca4b7635266372118c4ec1cde794aa72ace2bcb78fafe7f987ebf96ff4b350"} Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.280525 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bca4b7635266372118c4ec1cde794aa72ace2bcb78fafe7f987ebf96ff4b350" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.280309 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-zml5t" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.365116 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c"] Oct 04 11:12:53 crc kubenswrapper[4758]: E1004 11:12:53.365489 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7817b9e-b878-40b6-b1a5-9ed2526c32a5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.365508 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7817b9e-b878-40b6-b1a5-9ed2526c32a5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.365669 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7817b9e-b878-40b6-b1a5-9ed2526c32a5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.366323 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.368730 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.369128 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.369210 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.369452 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.383566 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c"] Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.538856 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.538979 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z7tv\" (UniqueName: \"kubernetes.io/projected/1b2357af-2674-4b82-b4fe-35ae6c2099d4-kube-api-access-9z7tv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.539013 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.539065 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.640996 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z7tv\" (UniqueName: \"kubernetes.io/projected/1b2357af-2674-4b82-b4fe-35ae6c2099d4-kube-api-access-9z7tv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.641359 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.641491 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.641620 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.647922 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.647951 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.648338 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.661584 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z7tv\" (UniqueName: \"kubernetes.io/projected/1b2357af-2674-4b82-b4fe-35ae6c2099d4-kube-api-access-9z7tv\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:53 crc kubenswrapper[4758]: I1004 11:12:53.686634 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:12:54 crc kubenswrapper[4758]: I1004 11:12:54.362403 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c"] Oct 04 11:12:55 crc kubenswrapper[4758]: I1004 11:12:55.346845 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" event={"ID":"1b2357af-2674-4b82-b4fe-35ae6c2099d4","Type":"ContainerStarted","Data":"f6ccec9af7e78fbd708f3fe036a65c2cd5ba2aaf864bab3741ccd4fd0aee7b2d"} Oct 04 11:12:55 crc kubenswrapper[4758]: I1004 11:12:55.348359 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" event={"ID":"1b2357af-2674-4b82-b4fe-35ae6c2099d4","Type":"ContainerStarted","Data":"4e9e19753418ce2b262f563a80c016d1e179230f524285b385f2cfe006284f29"} Oct 04 11:12:55 crc kubenswrapper[4758]: I1004 11:12:55.367572 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" podStartSLOduration=1.946645288 podStartE2EDuration="2.367515159s" podCreationTimestamp="2025-10-04 11:12:53 +0000 UTC" firstStartedPulling="2025-10-04 11:12:54.371916234 +0000 UTC m=+1351.664567133" lastFinishedPulling="2025-10-04 11:12:54.792786115 +0000 UTC m=+1352.085437004" observedRunningTime="2025-10-04 11:12:55.360723578 +0000 UTC m=+1352.653374487" watchObservedRunningTime="2025-10-04 11:12:55.367515159 +0000 UTC m=+1352.660166058" Oct 04 11:12:55 crc kubenswrapper[4758]: I1004 11:12:55.634755 4758 scope.go:117] "RemoveContainer" containerID="99c77c2e50490db809ce87a8529b681d0f7a43ec57af42e4eb421689a774e199" Oct 04 11:13:01 crc kubenswrapper[4758]: I1004 11:13:01.249761 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:13:01 crc kubenswrapper[4758]: I1004 11:13:01.250448 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:13:31 crc kubenswrapper[4758]: I1004 11:13:31.249618 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:13:31 crc kubenswrapper[4758]: I1004 11:13:31.250230 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:13:55 crc kubenswrapper[4758]: I1004 11:13:55.778123 4758 scope.go:117] "RemoveContainer" containerID="7a8a8092571e2e857f73ce667c8831efeffaa296cc5f54f0c38f852454d3790a" Oct 04 11:14:01 crc kubenswrapper[4758]: I1004 11:14:01.250011 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:14:01 crc kubenswrapper[4758]: I1004 11:14:01.251346 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:14:01 crc kubenswrapper[4758]: I1004 11:14:01.251440 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:14:01 crc kubenswrapper[4758]: I1004 11:14:01.252533 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"915ebcfb6b28055c5624cc8d51ba7691c189880fd98368198393745346dfbd64"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:14:01 crc kubenswrapper[4758]: I1004 11:14:01.252667 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://915ebcfb6b28055c5624cc8d51ba7691c189880fd98368198393745346dfbd64" gracePeriod=600 Oct 04 11:14:02 crc kubenswrapper[4758]: I1004 11:14:02.042362 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="915ebcfb6b28055c5624cc8d51ba7691c189880fd98368198393745346dfbd64" exitCode=0 Oct 04 11:14:02 crc kubenswrapper[4758]: I1004 11:14:02.042459 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"915ebcfb6b28055c5624cc8d51ba7691c189880fd98368198393745346dfbd64"} Oct 04 11:14:02 crc kubenswrapper[4758]: I1004 11:14:02.042925 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55"} Oct 04 11:14:02 crc kubenswrapper[4758]: I1004 11:14:02.042951 4758 scope.go:117] "RemoveContainer" containerID="56399f3880bb0482a5458063cd6de09cb352f18aa4d98a39db29016b0272c352" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.170197 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zwq65"] Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.175279 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.196972 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zwq65"] Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.240324 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-catalog-content\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.240532 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-utilities\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.240587 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjg9g\" (UniqueName: \"kubernetes.io/projected/027019d1-f398-42e2-a8cf-beedbb9513da-kube-api-access-mjg9g\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.342117 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-catalog-content\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.342212 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-utilities\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.342236 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjg9g\" (UniqueName: \"kubernetes.io/projected/027019d1-f398-42e2-a8cf-beedbb9513da-kube-api-access-mjg9g\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.342533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-catalog-content\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.342682 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-utilities\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.366966 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjg9g\" (UniqueName: \"kubernetes.io/projected/027019d1-f398-42e2-a8cf-beedbb9513da-kube-api-access-mjg9g\") pod \"certified-operators-zwq65\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.506328 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:48 crc kubenswrapper[4758]: I1004 11:14:48.992048 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zwq65"] Oct 04 11:14:49 crc kubenswrapper[4758]: I1004 11:14:49.577243 4758 generic.go:334] "Generic (PLEG): container finished" podID="027019d1-f398-42e2-a8cf-beedbb9513da" containerID="a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb" exitCode=0 Oct 04 11:14:49 crc kubenswrapper[4758]: I1004 11:14:49.577314 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerDied","Data":"a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb"} Oct 04 11:14:49 crc kubenswrapper[4758]: I1004 11:14:49.577675 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerStarted","Data":"6bdfa3e6e21d75280327eeb7ffa768e06df3f493d29381ceaa462571e0aa57c7"} Oct 04 11:14:50 crc kubenswrapper[4758]: I1004 11:14:50.589412 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerStarted","Data":"ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254"} Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.528206 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6g95r"] Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.530290 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.547529 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6g95r"] Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.605804 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-catalog-content\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.606003 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl5sf\" (UniqueName: \"kubernetes.io/projected/e1593324-052b-44ee-8aa8-56ddaa1bf645-kube-api-access-fl5sf\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.606029 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-utilities\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.707465 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl5sf\" (UniqueName: \"kubernetes.io/projected/e1593324-052b-44ee-8aa8-56ddaa1bf645-kube-api-access-fl5sf\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.707507 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-utilities\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.707558 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-catalog-content\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.708224 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-utilities\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.708336 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-catalog-content\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.727335 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl5sf\" (UniqueName: \"kubernetes.io/projected/e1593324-052b-44ee-8aa8-56ddaa1bf645-kube-api-access-fl5sf\") pod \"redhat-operators-6g95r\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:51 crc kubenswrapper[4758]: I1004 11:14:51.845524 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:14:52 crc kubenswrapper[4758]: I1004 11:14:52.323497 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6g95r"] Oct 04 11:14:52 crc kubenswrapper[4758]: I1004 11:14:52.607388 4758 generic.go:334] "Generic (PLEG): container finished" podID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerID="9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2" exitCode=0 Oct 04 11:14:52 crc kubenswrapper[4758]: I1004 11:14:52.607434 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerDied","Data":"9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2"} Oct 04 11:14:52 crc kubenswrapper[4758]: I1004 11:14:52.607477 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerStarted","Data":"1b083def97600b33aaf930f41ceb329f17ed12c8b4b5a07b5a7a954b9cf937d9"} Oct 04 11:14:52 crc kubenswrapper[4758]: I1004 11:14:52.609461 4758 generic.go:334] "Generic (PLEG): container finished" podID="027019d1-f398-42e2-a8cf-beedbb9513da" containerID="ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254" exitCode=0 Oct 04 11:14:52 crc kubenswrapper[4758]: I1004 11:14:52.609488 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerDied","Data":"ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254"} Oct 04 11:14:53 crc kubenswrapper[4758]: I1004 11:14:53.639697 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerStarted","Data":"f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa"} Oct 04 11:14:53 crc kubenswrapper[4758]: I1004 11:14:53.673311 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zwq65" podStartSLOduration=2.233441896 podStartE2EDuration="5.673286241s" podCreationTimestamp="2025-10-04 11:14:48 +0000 UTC" firstStartedPulling="2025-10-04 11:14:49.578484184 +0000 UTC m=+1466.871135083" lastFinishedPulling="2025-10-04 11:14:53.018328539 +0000 UTC m=+1470.310979428" observedRunningTime="2025-10-04 11:14:53.660288495 +0000 UTC m=+1470.952939384" watchObservedRunningTime="2025-10-04 11:14:53.673286241 +0000 UTC m=+1470.965937150" Oct 04 11:14:54 crc kubenswrapper[4758]: I1004 11:14:54.650774 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerStarted","Data":"01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c"} Oct 04 11:14:55 crc kubenswrapper[4758]: I1004 11:14:55.907242 4758 scope.go:117] "RemoveContainer" containerID="efa50282505e3dcde99e1070031fc0f6d76ac103d15def812dfbb6bcb67342c6" Oct 04 11:14:55 crc kubenswrapper[4758]: I1004 11:14:55.937891 4758 scope.go:117] "RemoveContainer" containerID="6f8f9aa39bf4e0f9918645575c022dce250e9ca8cf631068777b1e92c68b1fce" Oct 04 11:14:58 crc kubenswrapper[4758]: I1004 11:14:58.507349 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:58 crc kubenswrapper[4758]: I1004 11:14:58.507761 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:14:58 crc kubenswrapper[4758]: I1004 11:14:58.714195 4758 generic.go:334] "Generic (PLEG): container finished" podID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerID="01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c" exitCode=0 Oct 04 11:14:58 crc kubenswrapper[4758]: I1004 11:14:58.714260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerDied","Data":"01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c"} Oct 04 11:14:59 crc kubenswrapper[4758]: I1004 11:14:59.583971 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zwq65" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="registry-server" probeResult="failure" output=< Oct 04 11:14:59 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:14:59 crc kubenswrapper[4758]: > Oct 04 11:14:59 crc kubenswrapper[4758]: I1004 11:14:59.727365 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerStarted","Data":"fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d"} Oct 04 11:14:59 crc kubenswrapper[4758]: I1004 11:14:59.755347 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6g95r" podStartSLOduration=2.056697231 podStartE2EDuration="8.755325561s" podCreationTimestamp="2025-10-04 11:14:51 +0000 UTC" firstStartedPulling="2025-10-04 11:14:52.608937197 +0000 UTC m=+1469.901588086" lastFinishedPulling="2025-10-04 11:14:59.307565527 +0000 UTC m=+1476.600216416" observedRunningTime="2025-10-04 11:14:59.748327635 +0000 UTC m=+1477.040978534" watchObservedRunningTime="2025-10-04 11:14:59.755325561 +0000 UTC m=+1477.047976450" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.148051 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6"] Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.149691 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.155769 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.157215 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.161176 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6"] Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.270642 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/870c503b-cd34-4a52-959c-4b58bbeaf619-secret-volume\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.270857 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/870c503b-cd34-4a52-959c-4b58bbeaf619-config-volume\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.270909 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7567\" (UniqueName: \"kubernetes.io/projected/870c503b-cd34-4a52-959c-4b58bbeaf619-kube-api-access-c7567\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.374454 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7567\" (UniqueName: \"kubernetes.io/projected/870c503b-cd34-4a52-959c-4b58bbeaf619-kube-api-access-c7567\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.374576 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/870c503b-cd34-4a52-959c-4b58bbeaf619-secret-volume\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.374872 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/870c503b-cd34-4a52-959c-4b58bbeaf619-config-volume\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.384777 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/870c503b-cd34-4a52-959c-4b58bbeaf619-secret-volume\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.384960 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/870c503b-cd34-4a52-959c-4b58bbeaf619-config-volume\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.409409 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7567\" (UniqueName: \"kubernetes.io/projected/870c503b-cd34-4a52-959c-4b58bbeaf619-kube-api-access-c7567\") pod \"collect-profiles-29326275-rbpk6\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.469280 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:00 crc kubenswrapper[4758]: I1004 11:15:00.963458 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6"] Oct 04 11:15:01 crc kubenswrapper[4758]: I1004 11:15:01.744336 4758 generic.go:334] "Generic (PLEG): container finished" podID="870c503b-cd34-4a52-959c-4b58bbeaf619" containerID="723c9b921b1dd264326f3a7d78c5e59b30296488d42b758868322b893dfbeec7" exitCode=0 Oct 04 11:15:01 crc kubenswrapper[4758]: I1004 11:15:01.744492 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" event={"ID":"870c503b-cd34-4a52-959c-4b58bbeaf619","Type":"ContainerDied","Data":"723c9b921b1dd264326f3a7d78c5e59b30296488d42b758868322b893dfbeec7"} Oct 04 11:15:01 crc kubenswrapper[4758]: I1004 11:15:01.744703 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" event={"ID":"870c503b-cd34-4a52-959c-4b58bbeaf619","Type":"ContainerStarted","Data":"7cd30000f1aaab6a33dc12fde4d19f859cacdf5accafbea0435bd188c3d6a347"} Oct 04 11:15:01 crc kubenswrapper[4758]: I1004 11:15:01.846707 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:15:01 crc kubenswrapper[4758]: I1004 11:15:01.847397 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:15:02 crc kubenswrapper[4758]: I1004 11:15:02.900337 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6g95r" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" probeResult="failure" output=< Oct 04 11:15:02 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:15:02 crc kubenswrapper[4758]: > Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.121764 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.267633 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7567\" (UniqueName: \"kubernetes.io/projected/870c503b-cd34-4a52-959c-4b58bbeaf619-kube-api-access-c7567\") pod \"870c503b-cd34-4a52-959c-4b58bbeaf619\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.267804 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/870c503b-cd34-4a52-959c-4b58bbeaf619-secret-volume\") pod \"870c503b-cd34-4a52-959c-4b58bbeaf619\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.268605 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/870c503b-cd34-4a52-959c-4b58bbeaf619-config-volume\") pod \"870c503b-cd34-4a52-959c-4b58bbeaf619\" (UID: \"870c503b-cd34-4a52-959c-4b58bbeaf619\") " Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.269189 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870c503b-cd34-4a52-959c-4b58bbeaf619-config-volume" (OuterVolumeSpecName: "config-volume") pod "870c503b-cd34-4a52-959c-4b58bbeaf619" (UID: "870c503b-cd34-4a52-959c-4b58bbeaf619"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.287249 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/870c503b-cd34-4a52-959c-4b58bbeaf619-kube-api-access-c7567" (OuterVolumeSpecName: "kube-api-access-c7567") pod "870c503b-cd34-4a52-959c-4b58bbeaf619" (UID: "870c503b-cd34-4a52-959c-4b58bbeaf619"). InnerVolumeSpecName "kube-api-access-c7567". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.287371 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870c503b-cd34-4a52-959c-4b58bbeaf619-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "870c503b-cd34-4a52-959c-4b58bbeaf619" (UID: "870c503b-cd34-4a52-959c-4b58bbeaf619"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.370936 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7567\" (UniqueName: \"kubernetes.io/projected/870c503b-cd34-4a52-959c-4b58bbeaf619-kube-api-access-c7567\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.370979 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/870c503b-cd34-4a52-959c-4b58bbeaf619-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.370989 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/870c503b-cd34-4a52-959c-4b58bbeaf619-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.764121 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" event={"ID":"870c503b-cd34-4a52-959c-4b58bbeaf619","Type":"ContainerDied","Data":"7cd30000f1aaab6a33dc12fde4d19f859cacdf5accafbea0435bd188c3d6a347"} Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.764466 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cd30000f1aaab6a33dc12fde4d19f859cacdf5accafbea0435bd188c3d6a347" Oct 04 11:15:03 crc kubenswrapper[4758]: I1004 11:15:03.764215 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6" Oct 04 11:15:08 crc kubenswrapper[4758]: I1004 11:15:08.557667 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:15:08 crc kubenswrapper[4758]: I1004 11:15:08.610366 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:15:08 crc kubenswrapper[4758]: I1004 11:15:08.794399 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zwq65"] Oct 04 11:15:09 crc kubenswrapper[4758]: I1004 11:15:09.820400 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zwq65" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="registry-server" containerID="cri-o://f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa" gracePeriod=2 Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.259404 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.412094 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-utilities\") pod \"027019d1-f398-42e2-a8cf-beedbb9513da\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.412213 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-catalog-content\") pod \"027019d1-f398-42e2-a8cf-beedbb9513da\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.412365 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjg9g\" (UniqueName: \"kubernetes.io/projected/027019d1-f398-42e2-a8cf-beedbb9513da-kube-api-access-mjg9g\") pod \"027019d1-f398-42e2-a8cf-beedbb9513da\" (UID: \"027019d1-f398-42e2-a8cf-beedbb9513da\") " Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.414728 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-utilities" (OuterVolumeSpecName: "utilities") pod "027019d1-f398-42e2-a8cf-beedbb9513da" (UID: "027019d1-f398-42e2-a8cf-beedbb9513da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.418822 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/027019d1-f398-42e2-a8cf-beedbb9513da-kube-api-access-mjg9g" (OuterVolumeSpecName: "kube-api-access-mjg9g") pod "027019d1-f398-42e2-a8cf-beedbb9513da" (UID: "027019d1-f398-42e2-a8cf-beedbb9513da"). InnerVolumeSpecName "kube-api-access-mjg9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.465354 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "027019d1-f398-42e2-a8cf-beedbb9513da" (UID: "027019d1-f398-42e2-a8cf-beedbb9513da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.514229 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.514261 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjg9g\" (UniqueName: \"kubernetes.io/projected/027019d1-f398-42e2-a8cf-beedbb9513da-kube-api-access-mjg9g\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.514273 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/027019d1-f398-42e2-a8cf-beedbb9513da-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.835249 4758 generic.go:334] "Generic (PLEG): container finished" podID="027019d1-f398-42e2-a8cf-beedbb9513da" containerID="f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa" exitCode=0 Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.835322 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zwq65" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.835327 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerDied","Data":"f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa"} Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.835513 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zwq65" event={"ID":"027019d1-f398-42e2-a8cf-beedbb9513da","Type":"ContainerDied","Data":"6bdfa3e6e21d75280327eeb7ffa768e06df3f493d29381ceaa462571e0aa57c7"} Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.835558 4758 scope.go:117] "RemoveContainer" containerID="f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.870606 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zwq65"] Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.877362 4758 scope.go:117] "RemoveContainer" containerID="ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.879152 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zwq65"] Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.920651 4758 scope.go:117] "RemoveContainer" containerID="a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.965374 4758 scope.go:117] "RemoveContainer" containerID="f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa" Oct 04 11:15:10 crc kubenswrapper[4758]: E1004 11:15:10.965812 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa\": container with ID starting with f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa not found: ID does not exist" containerID="f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.965900 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa"} err="failed to get container status \"f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa\": rpc error: code = NotFound desc = could not find container \"f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa\": container with ID starting with f41c4cb2c021f10afbbd91cfcb0fc2ce72976c382a64421fef74362fdc6902aa not found: ID does not exist" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.965932 4758 scope.go:117] "RemoveContainer" containerID="ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254" Oct 04 11:15:10 crc kubenswrapper[4758]: E1004 11:15:10.966198 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254\": container with ID starting with ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254 not found: ID does not exist" containerID="ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.966231 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254"} err="failed to get container status \"ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254\": rpc error: code = NotFound desc = could not find container \"ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254\": container with ID starting with ccbc51e37076ad81ad7212cd055bc45dd5adc473ce63c79fdc5d1debc422e254 not found: ID does not exist" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.966249 4758 scope.go:117] "RemoveContainer" containerID="a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb" Oct 04 11:15:10 crc kubenswrapper[4758]: E1004 11:15:10.966524 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb\": container with ID starting with a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb not found: ID does not exist" containerID="a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb" Oct 04 11:15:10 crc kubenswrapper[4758]: I1004 11:15:10.966557 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb"} err="failed to get container status \"a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb\": rpc error: code = NotFound desc = could not find container \"a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb\": container with ID starting with a44cfe7273099f4c23e00012d87055b6ba49bfb69ebebc70b9510c635ba0b7cb not found: ID does not exist" Oct 04 11:15:11 crc kubenswrapper[4758]: I1004 11:15:11.337610 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" path="/var/lib/kubelet/pods/027019d1-f398-42e2-a8cf-beedbb9513da/volumes" Oct 04 11:15:12 crc kubenswrapper[4758]: I1004 11:15:12.890667 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6g95r" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" probeResult="failure" output=< Oct 04 11:15:12 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:15:12 crc kubenswrapper[4758]: > Oct 04 11:15:22 crc kubenswrapper[4758]: I1004 11:15:22.896440 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6g95r" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" probeResult="failure" output=< Oct 04 11:15:22 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:15:22 crc kubenswrapper[4758]: > Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.143423 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5mtnn"] Oct 04 11:15:25 crc kubenswrapper[4758]: E1004 11:15:25.144688 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="registry-server" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.144713 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="registry-server" Oct 04 11:15:25 crc kubenswrapper[4758]: E1004 11:15:25.144746 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="870c503b-cd34-4a52-959c-4b58bbeaf619" containerName="collect-profiles" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.144759 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="870c503b-cd34-4a52-959c-4b58bbeaf619" containerName="collect-profiles" Oct 04 11:15:25 crc kubenswrapper[4758]: E1004 11:15:25.144794 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="extract-utilities" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.144808 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="extract-utilities" Oct 04 11:15:25 crc kubenswrapper[4758]: E1004 11:15:25.144824 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="extract-content" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.144836 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="extract-content" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.145176 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="027019d1-f398-42e2-a8cf-beedbb9513da" containerName="registry-server" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.145206 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="870c503b-cd34-4a52-959c-4b58bbeaf619" containerName="collect-profiles" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.147899 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.162064 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5mtnn"] Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.213028 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5rfl\" (UniqueName: \"kubernetes.io/projected/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-kube-api-access-b5rfl\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.213154 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-catalog-content\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.213195 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-utilities\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.314588 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5rfl\" (UniqueName: \"kubernetes.io/projected/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-kube-api-access-b5rfl\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.314675 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-catalog-content\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.314700 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-utilities\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.315211 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-utilities\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.315283 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-catalog-content\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.337336 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5rfl\" (UniqueName: \"kubernetes.io/projected/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-kube-api-access-b5rfl\") pod \"community-operators-5mtnn\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.477774 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.941213 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5mtnn"] Oct 04 11:15:25 crc kubenswrapper[4758]: I1004 11:15:25.970712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerStarted","Data":"3b54ffc72a375c556e33dee08eed5117165cc47a35ae0175d687c55663020847"} Oct 04 11:15:26 crc kubenswrapper[4758]: I1004 11:15:26.982480 4758 generic.go:334] "Generic (PLEG): container finished" podID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerID="4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4" exitCode=0 Oct 04 11:15:26 crc kubenswrapper[4758]: I1004 11:15:26.982680 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerDied","Data":"4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4"} Oct 04 11:15:29 crc kubenswrapper[4758]: I1004 11:15:29.005335 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerStarted","Data":"cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e"} Oct 04 11:15:30 crc kubenswrapper[4758]: I1004 11:15:30.016010 4758 generic.go:334] "Generic (PLEG): container finished" podID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerID="cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e" exitCode=0 Oct 04 11:15:30 crc kubenswrapper[4758]: I1004 11:15:30.016222 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerDied","Data":"cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e"} Oct 04 11:15:31 crc kubenswrapper[4758]: I1004 11:15:31.027072 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerStarted","Data":"3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134"} Oct 04 11:15:31 crc kubenswrapper[4758]: I1004 11:15:31.052568 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5mtnn" podStartSLOduration=2.48306455 podStartE2EDuration="6.052551048s" podCreationTimestamp="2025-10-04 11:15:25 +0000 UTC" firstStartedPulling="2025-10-04 11:15:26.984672707 +0000 UTC m=+1504.277323596" lastFinishedPulling="2025-10-04 11:15:30.554159205 +0000 UTC m=+1507.846810094" observedRunningTime="2025-10-04 11:15:31.045819429 +0000 UTC m=+1508.338470318" watchObservedRunningTime="2025-10-04 11:15:31.052551048 +0000 UTC m=+1508.345201937" Oct 04 11:15:31 crc kubenswrapper[4758]: I1004 11:15:31.903172 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:15:31 crc kubenswrapper[4758]: I1004 11:15:31.957077 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:15:32 crc kubenswrapper[4758]: I1004 11:15:32.494497 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6g95r"] Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.056658 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6g95r" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" containerID="cri-o://fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d" gracePeriod=2 Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.472446 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.566393 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-catalog-content\") pod \"e1593324-052b-44ee-8aa8-56ddaa1bf645\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.566457 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl5sf\" (UniqueName: \"kubernetes.io/projected/e1593324-052b-44ee-8aa8-56ddaa1bf645-kube-api-access-fl5sf\") pod \"e1593324-052b-44ee-8aa8-56ddaa1bf645\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.566531 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-utilities\") pod \"e1593324-052b-44ee-8aa8-56ddaa1bf645\" (UID: \"e1593324-052b-44ee-8aa8-56ddaa1bf645\") " Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.567499 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-utilities" (OuterVolumeSpecName: "utilities") pod "e1593324-052b-44ee-8aa8-56ddaa1bf645" (UID: "e1593324-052b-44ee-8aa8-56ddaa1bf645"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.585268 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1593324-052b-44ee-8aa8-56ddaa1bf645-kube-api-access-fl5sf" (OuterVolumeSpecName: "kube-api-access-fl5sf") pod "e1593324-052b-44ee-8aa8-56ddaa1bf645" (UID: "e1593324-052b-44ee-8aa8-56ddaa1bf645"). InnerVolumeSpecName "kube-api-access-fl5sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.669124 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl5sf\" (UniqueName: \"kubernetes.io/projected/e1593324-052b-44ee-8aa8-56ddaa1bf645-kube-api-access-fl5sf\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.669162 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.677338 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1593324-052b-44ee-8aa8-56ddaa1bf645" (UID: "e1593324-052b-44ee-8aa8-56ddaa1bf645"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:15:33 crc kubenswrapper[4758]: I1004 11:15:33.770454 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1593324-052b-44ee-8aa8-56ddaa1bf645-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.066554 4758 generic.go:334] "Generic (PLEG): container finished" podID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerID="fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d" exitCode=0 Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.066597 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerDied","Data":"fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d"} Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.066631 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6g95r" event={"ID":"e1593324-052b-44ee-8aa8-56ddaa1bf645","Type":"ContainerDied","Data":"1b083def97600b33aaf930f41ceb329f17ed12c8b4b5a07b5a7a954b9cf937d9"} Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.066635 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6g95r" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.066655 4758 scope.go:117] "RemoveContainer" containerID="fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.084273 4758 scope.go:117] "RemoveContainer" containerID="01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.106566 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6g95r"] Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.121724 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6g95r"] Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.124272 4758 scope.go:117] "RemoveContainer" containerID="9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.149533 4758 scope.go:117] "RemoveContainer" containerID="fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d" Oct 04 11:15:34 crc kubenswrapper[4758]: E1004 11:15:34.149943 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d\": container with ID starting with fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d not found: ID does not exist" containerID="fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.149967 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d"} err="failed to get container status \"fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d\": rpc error: code = NotFound desc = could not find container \"fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d\": container with ID starting with fae86325338506342bbf1619fa11ebb00f15014e545312135d5983ccd74e931d not found: ID does not exist" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.149987 4758 scope.go:117] "RemoveContainer" containerID="01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c" Oct 04 11:15:34 crc kubenswrapper[4758]: E1004 11:15:34.150180 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c\": container with ID starting with 01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c not found: ID does not exist" containerID="01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.150197 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c"} err="failed to get container status \"01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c\": rpc error: code = NotFound desc = could not find container \"01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c\": container with ID starting with 01d763e59915dac58e439c3c081d6c5edec4868f8c1a324a29d4fbb346710a9c not found: ID does not exist" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.150209 4758 scope.go:117] "RemoveContainer" containerID="9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2" Oct 04 11:15:34 crc kubenswrapper[4758]: E1004 11:15:34.150459 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2\": container with ID starting with 9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2 not found: ID does not exist" containerID="9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2" Oct 04 11:15:34 crc kubenswrapper[4758]: I1004 11:15:34.150476 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2"} err="failed to get container status \"9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2\": rpc error: code = NotFound desc = could not find container \"9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2\": container with ID starting with 9f433498e5aecfdc0c3859e6354d333fc6050dc356db5145a8857d5f5e9afbb2 not found: ID does not exist" Oct 04 11:15:35 crc kubenswrapper[4758]: I1004 11:15:35.337068 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" path="/var/lib/kubelet/pods/e1593324-052b-44ee-8aa8-56ddaa1bf645/volumes" Oct 04 11:15:35 crc kubenswrapper[4758]: I1004 11:15:35.478517 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:35 crc kubenswrapper[4758]: I1004 11:15:35.478599 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:35 crc kubenswrapper[4758]: I1004 11:15:35.533488 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:36 crc kubenswrapper[4758]: I1004 11:15:36.132383 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:36 crc kubenswrapper[4758]: I1004 11:15:36.891882 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5mtnn"] Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.106326 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5mtnn" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="registry-server" containerID="cri-o://3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134" gracePeriod=2 Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.581730 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.653209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5rfl\" (UniqueName: \"kubernetes.io/projected/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-kube-api-access-b5rfl\") pod \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.653552 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-catalog-content\") pod \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.653669 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-utilities\") pod \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\" (UID: \"be544bfd-56b3-41c9-8d6c-a4f9496d86bc\") " Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.654321 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-utilities" (OuterVolumeSpecName: "utilities") pod "be544bfd-56b3-41c9-8d6c-a4f9496d86bc" (UID: "be544bfd-56b3-41c9-8d6c-a4f9496d86bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.673551 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-kube-api-access-b5rfl" (OuterVolumeSpecName: "kube-api-access-b5rfl") pod "be544bfd-56b3-41c9-8d6c-a4f9496d86bc" (UID: "be544bfd-56b3-41c9-8d6c-a4f9496d86bc"). InnerVolumeSpecName "kube-api-access-b5rfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.714533 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be544bfd-56b3-41c9-8d6c-a4f9496d86bc" (UID: "be544bfd-56b3-41c9-8d6c-a4f9496d86bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.756681 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.756901 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:38 crc kubenswrapper[4758]: I1004 11:15:38.756958 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5rfl\" (UniqueName: \"kubernetes.io/projected/be544bfd-56b3-41c9-8d6c-a4f9496d86bc-kube-api-access-b5rfl\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.115128 4758 generic.go:334] "Generic (PLEG): container finished" podID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerID="3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134" exitCode=0 Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.115180 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5mtnn" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.115196 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerDied","Data":"3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134"} Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.116478 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5mtnn" event={"ID":"be544bfd-56b3-41c9-8d6c-a4f9496d86bc","Type":"ContainerDied","Data":"3b54ffc72a375c556e33dee08eed5117165cc47a35ae0175d687c55663020847"} Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.116499 4758 scope.go:117] "RemoveContainer" containerID="3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.151452 4758 scope.go:117] "RemoveContainer" containerID="cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.179156 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5mtnn"] Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.254893 4758 scope.go:117] "RemoveContainer" containerID="4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.264532 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5mtnn"] Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.333333 4758 scope.go:117] "RemoveContainer" containerID="3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134" Oct 04 11:15:39 crc kubenswrapper[4758]: E1004 11:15:39.333853 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134\": container with ID starting with 3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134 not found: ID does not exist" containerID="3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.333901 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134"} err="failed to get container status \"3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134\": rpc error: code = NotFound desc = could not find container \"3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134\": container with ID starting with 3513fe32ebad32431d87510f1e36ea5b60f5b32511a3cc0509b4f8a6d8137134 not found: ID does not exist" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.333929 4758 scope.go:117] "RemoveContainer" containerID="cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e" Oct 04 11:15:39 crc kubenswrapper[4758]: E1004 11:15:39.334202 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e\": container with ID starting with cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e not found: ID does not exist" containerID="cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.334242 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e"} err="failed to get container status \"cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e\": rpc error: code = NotFound desc = could not find container \"cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e\": container with ID starting with cc20d03c113a889c8c5038f532c63f2666476f67fc5e45d95ad36f8bf041ae5e not found: ID does not exist" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.334264 4758 scope.go:117] "RemoveContainer" containerID="4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4" Oct 04 11:15:39 crc kubenswrapper[4758]: E1004 11:15:39.334503 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4\": container with ID starting with 4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4 not found: ID does not exist" containerID="4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.334537 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4"} err="failed to get container status \"4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4\": rpc error: code = NotFound desc = could not find container \"4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4\": container with ID starting with 4fe890b7e84b4d5a4bebe0f9f38297c6abb68033b8c4215263ecf2fecefe9de4 not found: ID does not exist" Oct 04 11:15:39 crc kubenswrapper[4758]: I1004 11:15:39.346251 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" path="/var/lib/kubelet/pods/be544bfd-56b3-41c9-8d6c-a4f9496d86bc/volumes" Oct 04 11:15:55 crc kubenswrapper[4758]: I1004 11:15:55.983260 4758 scope.go:117] "RemoveContainer" containerID="d79da546ff3f3946d480c3b033c65d953b200d453e8aae9a6eb7e5b1ab370dfd" Oct 04 11:15:56 crc kubenswrapper[4758]: I1004 11:15:56.011491 4758 scope.go:117] "RemoveContainer" containerID="9209fc34d8a144b1d61f51ddfba3f65281665b1eb154215091660ba780d10b2e" Oct 04 11:16:01 crc kubenswrapper[4758]: I1004 11:16:01.249574 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:16:01 crc kubenswrapper[4758]: I1004 11:16:01.250282 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:16:10 crc kubenswrapper[4758]: I1004 11:16:10.403215 4758 generic.go:334] "Generic (PLEG): container finished" podID="1b2357af-2674-4b82-b4fe-35ae6c2099d4" containerID="f6ccec9af7e78fbd708f3fe036a65c2cd5ba2aaf864bab3741ccd4fd0aee7b2d" exitCode=0 Oct 04 11:16:10 crc kubenswrapper[4758]: I1004 11:16:10.403275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" event={"ID":"1b2357af-2674-4b82-b4fe-35ae6c2099d4","Type":"ContainerDied","Data":"f6ccec9af7e78fbd708f3fe036a65c2cd5ba2aaf864bab3741ccd4fd0aee7b2d"} Oct 04 11:16:11 crc kubenswrapper[4758]: I1004 11:16:11.862354 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.023788 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z7tv\" (UniqueName: \"kubernetes.io/projected/1b2357af-2674-4b82-b4fe-35ae6c2099d4-kube-api-access-9z7tv\") pod \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.023933 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-bootstrap-combined-ca-bundle\") pod \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.023968 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-ssh-key\") pod \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.023991 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-inventory\") pod \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\" (UID: \"1b2357af-2674-4b82-b4fe-35ae6c2099d4\") " Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.030316 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1b2357af-2674-4b82-b4fe-35ae6c2099d4" (UID: "1b2357af-2674-4b82-b4fe-35ae6c2099d4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.030470 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2357af-2674-4b82-b4fe-35ae6c2099d4-kube-api-access-9z7tv" (OuterVolumeSpecName: "kube-api-access-9z7tv") pod "1b2357af-2674-4b82-b4fe-35ae6c2099d4" (UID: "1b2357af-2674-4b82-b4fe-35ae6c2099d4"). InnerVolumeSpecName "kube-api-access-9z7tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.054238 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-inventory" (OuterVolumeSpecName: "inventory") pod "1b2357af-2674-4b82-b4fe-35ae6c2099d4" (UID: "1b2357af-2674-4b82-b4fe-35ae6c2099d4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.055113 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b2357af-2674-4b82-b4fe-35ae6c2099d4" (UID: "1b2357af-2674-4b82-b4fe-35ae6c2099d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.126734 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z7tv\" (UniqueName: \"kubernetes.io/projected/1b2357af-2674-4b82-b4fe-35ae6c2099d4-kube-api-access-9z7tv\") on node \"crc\" DevicePath \"\"" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.126917 4758 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.126981 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.127039 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b2357af-2674-4b82-b4fe-35ae6c2099d4-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.422310 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" event={"ID":"1b2357af-2674-4b82-b4fe-35ae6c2099d4","Type":"ContainerDied","Data":"4e9e19753418ce2b262f563a80c016d1e179230f524285b385f2cfe006284f29"} Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.422346 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e9e19753418ce2b262f563a80c016d1e179230f524285b385f2cfe006284f29" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.422368 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.521900 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft"] Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.522522 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="extract-utilities" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.522596 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="extract-utilities" Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.522660 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="extract-content" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.522716 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="extract-content" Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.522776 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.522832 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.522891 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="extract-content" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.522951 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="extract-content" Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.523219 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2357af-2674-4b82-b4fe-35ae6c2099d4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.523278 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2357af-2674-4b82-b4fe-35ae6c2099d4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.523342 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="registry-server" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.523396 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="registry-server" Oct 04 11:16:12 crc kubenswrapper[4758]: E1004 11:16:12.523456 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="extract-utilities" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.523507 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="extract-utilities" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.523709 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2357af-2674-4b82-b4fe-35ae6c2099d4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.523791 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1593324-052b-44ee-8aa8-56ddaa1bf645" containerName="registry-server" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.523853 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="be544bfd-56b3-41c9-8d6c-a4f9496d86bc" containerName="registry-server" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.524595 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.526837 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.527375 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.527677 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.527848 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.541448 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft"] Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.635667 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.636012 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.636342 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j2ws\" (UniqueName: \"kubernetes.io/projected/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-kube-api-access-7j2ws\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.738437 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j2ws\" (UniqueName: \"kubernetes.io/projected/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-kube-api-access-7j2ws\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.738573 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.738690 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.742858 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.742892 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.758238 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j2ws\" (UniqueName: \"kubernetes.io/projected/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-kube-api-access-7j2ws\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-slrft\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:12 crc kubenswrapper[4758]: I1004 11:16:12.845232 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:16:13 crc kubenswrapper[4758]: I1004 11:16:13.417647 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft"] Oct 04 11:16:13 crc kubenswrapper[4758]: I1004 11:16:13.423853 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:16:13 crc kubenswrapper[4758]: I1004 11:16:13.430984 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" event={"ID":"ceea8611-3691-4c29-b4ae-5c1e0d0bb396","Type":"ContainerStarted","Data":"6e042a5a3057defece034649f77d2876984d11adbd3fba87452dbe74898621c2"} Oct 04 11:16:14 crc kubenswrapper[4758]: I1004 11:16:14.442489 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" event={"ID":"ceea8611-3691-4c29-b4ae-5c1e0d0bb396","Type":"ContainerStarted","Data":"ef246d3f0c5b1c6511eb2cf6c684b5de07bd7709425052e3cbeec76468a5cfee"} Oct 04 11:16:14 crc kubenswrapper[4758]: I1004 11:16:14.472737 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" podStartSLOduration=1.886769112 podStartE2EDuration="2.472723246s" podCreationTimestamp="2025-10-04 11:16:12 +0000 UTC" firstStartedPulling="2025-10-04 11:16:13.423653509 +0000 UTC m=+1550.716304398" lastFinishedPulling="2025-10-04 11:16:14.009607603 +0000 UTC m=+1551.302258532" observedRunningTime="2025-10-04 11:16:14.465766531 +0000 UTC m=+1551.758417420" watchObservedRunningTime="2025-10-04 11:16:14.472723246 +0000 UTC m=+1551.765374125" Oct 04 11:16:17 crc kubenswrapper[4758]: I1004 11:16:17.061791 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-z6vwk"] Oct 04 11:16:17 crc kubenswrapper[4758]: I1004 11:16:17.073148 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-68x7w"] Oct 04 11:16:17 crc kubenswrapper[4758]: I1004 11:16:17.081712 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-68x7w"] Oct 04 11:16:17 crc kubenswrapper[4758]: I1004 11:16:17.090497 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-z6vwk"] Oct 04 11:16:17 crc kubenswrapper[4758]: I1004 11:16:17.339813 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b43b9d-d8b6-45e8-83ae-432775163641" path="/var/lib/kubelet/pods/66b43b9d-d8b6-45e8-83ae-432775163641/volumes" Oct 04 11:16:17 crc kubenswrapper[4758]: I1004 11:16:17.346515 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f88e571-0f7b-4740-a4e2-3ce25c94bac9" path="/var/lib/kubelet/pods/6f88e571-0f7b-4740-a4e2-3ce25c94bac9/volumes" Oct 04 11:16:18 crc kubenswrapper[4758]: I1004 11:16:18.043579 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ktlf7"] Oct 04 11:16:18 crc kubenswrapper[4758]: I1004 11:16:18.053502 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ktlf7"] Oct 04 11:16:19 crc kubenswrapper[4758]: I1004 11:16:19.344709 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b59f374-8731-40e2-b34d-4874ba393094" path="/var/lib/kubelet/pods/5b59f374-8731-40e2-b34d-4874ba393094/volumes" Oct 04 11:16:27 crc kubenswrapper[4758]: I1004 11:16:27.032391 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e50b-account-create-vl6ml"] Oct 04 11:16:27 crc kubenswrapper[4758]: I1004 11:16:27.044625 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-80cb-account-create-qcgm2"] Oct 04 11:16:27 crc kubenswrapper[4758]: I1004 11:16:27.052046 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-80cb-account-create-qcgm2"] Oct 04 11:16:27 crc kubenswrapper[4758]: I1004 11:16:27.059397 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e50b-account-create-vl6ml"] Oct 04 11:16:27 crc kubenswrapper[4758]: I1004 11:16:27.336071 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8849e0bd-1093-4779-a8ae-16ded0d8622e" path="/var/lib/kubelet/pods/8849e0bd-1093-4779-a8ae-16ded0d8622e/volumes" Oct 04 11:16:27 crc kubenswrapper[4758]: I1004 11:16:27.338660 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c751e0-cd94-4006-8492-271f1125f266" path="/var/lib/kubelet/pods/c3c751e0-cd94-4006-8492-271f1125f266/volumes" Oct 04 11:16:31 crc kubenswrapper[4758]: I1004 11:16:31.249990 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:16:31 crc kubenswrapper[4758]: I1004 11:16:31.250624 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:16:32 crc kubenswrapper[4758]: I1004 11:16:32.028050 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-76e3-account-create-xn8n5"] Oct 04 11:16:32 crc kubenswrapper[4758]: I1004 11:16:32.035967 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-76e3-account-create-xn8n5"] Oct 04 11:16:33 crc kubenswrapper[4758]: I1004 11:16:33.339497 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48a3538c-a18f-41b2-87f6-59240be0d9f8" path="/var/lib/kubelet/pods/48a3538c-a18f-41b2-87f6-59240be0d9f8/volumes" Oct 04 11:16:34 crc kubenswrapper[4758]: I1004 11:16:34.029518 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8s4md"] Oct 04 11:16:34 crc kubenswrapper[4758]: I1004 11:16:34.037036 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-t5kz9"] Oct 04 11:16:34 crc kubenswrapper[4758]: I1004 11:16:34.044953 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8s4md"] Oct 04 11:16:34 crc kubenswrapper[4758]: I1004 11:16:34.051911 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-t5kz9"] Oct 04 11:16:35 crc kubenswrapper[4758]: I1004 11:16:35.024573 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-bgthn"] Oct 04 11:16:35 crc kubenswrapper[4758]: I1004 11:16:35.034090 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-bgthn"] Oct 04 11:16:35 crc kubenswrapper[4758]: I1004 11:16:35.339652 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="246cc3c0-2cb3-4def-bac8-54aa56554813" path="/var/lib/kubelet/pods/246cc3c0-2cb3-4def-bac8-54aa56554813/volumes" Oct 04 11:16:35 crc kubenswrapper[4758]: I1004 11:16:35.343600 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40ad036-101f-4e20-b049-cd53739e957f" path="/var/lib/kubelet/pods/c40ad036-101f-4e20-b049-cd53739e957f/volumes" Oct 04 11:16:35 crc kubenswrapper[4758]: I1004 11:16:35.346673 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c573ca52-55dd-4efd-9227-35d71b1f5268" path="/var/lib/kubelet/pods/c573ca52-55dd-4efd-9227-35d71b1f5268/volumes" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.110925 4758 scope.go:117] "RemoveContainer" containerID="dbcfb67bfdf15825765f735285dd714869298a659ccc5fd8587c1ca1ec10827b" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.153689 4758 scope.go:117] "RemoveContainer" containerID="6023cda852b544c11a75f84fcbe94b1fc1752d7b23f990722a98f3224c951939" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.180465 4758 scope.go:117] "RemoveContainer" containerID="7e0830d3b36673987fc3a4abd269a946c10ac75eabb8e0f19fba2aca12f5cae8" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.235017 4758 scope.go:117] "RemoveContainer" containerID="07329e5ec7c1aa1806f147ad17c098c9b755affa9b5b56374e764083d2023e85" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.266080 4758 scope.go:117] "RemoveContainer" containerID="cf3bbd33dc6d0dff561e88c7fbae037cd665c7080d779b7c10bae31d0b09175b" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.312174 4758 scope.go:117] "RemoveContainer" containerID="f0c1585415240feabf438e7f5a3d3c0965720f7d757687072f27536d4f50308d" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.363072 4758 scope.go:117] "RemoveContainer" containerID="9b164ac84dbd8ab507cbdea2656f78d8cbe32079002f4b0075d67b399025af07" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.402250 4758 scope.go:117] "RemoveContainer" containerID="85fff0c1ee73d1a1c33f5c599c71c593ae99077e09c4e0315b8736dfa9fa4b40" Oct 04 11:16:56 crc kubenswrapper[4758]: I1004 11:16:56.426787 4758 scope.go:117] "RemoveContainer" containerID="d2bb6e8e87efc9456c81cd214b0a3f2b4f2f2aba2969b1474ca5b45b78bffd85" Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.053455 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-bbb3-account-create-2frqn"] Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.064925 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6ad0-account-create-6rh8j"] Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.072795 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-4285-account-create-rzjvs"] Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.081836 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-4285-account-create-rzjvs"] Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.088981 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-bbb3-account-create-2frqn"] Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.095942 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6ad0-account-create-6rh8j"] Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.345473 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a3ed6de-8959-441c-bf24-025f5e7c185d" path="/var/lib/kubelet/pods/0a3ed6de-8959-441c-bf24-025f5e7c185d/volumes" Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.349603 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4" path="/var/lib/kubelet/pods/7fa5cd7e-8df4-42ad-bf1d-10728e3fb2c4/volumes" Oct 04 11:16:57 crc kubenswrapper[4758]: I1004 11:16:57.354053 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab228585-a058-4d2a-aac4-6dd6bbf30852" path="/var/lib/kubelet/pods/ab228585-a058-4d2a-aac4-6dd6bbf30852/volumes" Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.249164 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.249365 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.249473 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.250380 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.250478 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" gracePeriod=600 Oct 04 11:17:01 crc kubenswrapper[4758]: E1004 11:17:01.384427 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.895280 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" exitCode=0 Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.895319 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55"} Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.895348 4758 scope.go:117] "RemoveContainer" containerID="915ebcfb6b28055c5624cc8d51ba7691c189880fd98368198393745346dfbd64" Oct 04 11:17:01 crc kubenswrapper[4758]: I1004 11:17:01.895983 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:17:01 crc kubenswrapper[4758]: E1004 11:17:01.896224 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:12 crc kubenswrapper[4758]: I1004 11:17:12.327287 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:17:12 crc kubenswrapper[4758]: E1004 11:17:12.328657 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:13 crc kubenswrapper[4758]: I1004 11:17:13.060658 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-crlgt"] Oct 04 11:17:13 crc kubenswrapper[4758]: I1004 11:17:13.076664 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-crlgt"] Oct 04 11:17:13 crc kubenswrapper[4758]: I1004 11:17:13.336209 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="157f102f-c0ec-4e03-97cb-9b45948c82ea" path="/var/lib/kubelet/pods/157f102f-c0ec-4e03-97cb-9b45948c82ea/volumes" Oct 04 11:17:19 crc kubenswrapper[4758]: I1004 11:17:19.028164 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-trp6p"] Oct 04 11:17:19 crc kubenswrapper[4758]: I1004 11:17:19.039075 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-trp6p"] Oct 04 11:17:19 crc kubenswrapper[4758]: I1004 11:17:19.339605 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe10db6-4338-48ab-b744-18213468c240" path="/var/lib/kubelet/pods/bfe10db6-4338-48ab-b744-18213468c240/volumes" Oct 04 11:17:23 crc kubenswrapper[4758]: I1004 11:17:23.331540 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:17:23 crc kubenswrapper[4758]: E1004 11:17:23.332008 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:35 crc kubenswrapper[4758]: I1004 11:17:35.326118 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:17:35 crc kubenswrapper[4758]: E1004 11:17:35.326727 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:48 crc kubenswrapper[4758]: I1004 11:17:48.325654 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:17:48 crc kubenswrapper[4758]: E1004 11:17:48.326982 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:50 crc kubenswrapper[4758]: I1004 11:17:50.050612 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qqgl8"] Oct 04 11:17:50 crc kubenswrapper[4758]: I1004 11:17:50.061057 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qqgl8"] Oct 04 11:17:51 crc kubenswrapper[4758]: I1004 11:17:51.338454 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="707c36c5-c69b-4d1d-8a0a-43891279b487" path="/var/lib/kubelet/pods/707c36c5-c69b-4d1d-8a0a-43891279b487/volumes" Oct 04 11:17:56 crc kubenswrapper[4758]: I1004 11:17:56.649183 4758 scope.go:117] "RemoveContainer" containerID="646b378ed44240903bc73e92a9185804f43c84628a92aa4752dac7debdc99783" Oct 04 11:17:56 crc kubenswrapper[4758]: I1004 11:17:56.689281 4758 scope.go:117] "RemoveContainer" containerID="5bc7ddabb624c50b1415f39497cc2b97cebbd2598f9bd3e15deb94dc9ff29879" Oct 04 11:17:56 crc kubenswrapper[4758]: I1004 11:17:56.757357 4758 scope.go:117] "RemoveContainer" containerID="5d5c2ec209501eacb40bda392aafacf21398d2ec28df229e6b6b39fa3fd21693" Oct 04 11:17:56 crc kubenswrapper[4758]: I1004 11:17:56.837462 4758 scope.go:117] "RemoveContainer" containerID="880004199ec2201e3d172dc100ed19d53d3e62e0e2b829dd48097b64ba29c9b8" Oct 04 11:17:56 crc kubenswrapper[4758]: I1004 11:17:56.869736 4758 scope.go:117] "RemoveContainer" containerID="3d9139b91b63076052894750d8ed65555d0b934b83aeecf48a84686a4b61d02e" Oct 04 11:17:56 crc kubenswrapper[4758]: I1004 11:17:56.930725 4758 scope.go:117] "RemoveContainer" containerID="733a2dbe8d4aa5ac353094a971ab1255ccba823e4bff633aaaaaea5e0500922b" Oct 04 11:17:59 crc kubenswrapper[4758]: I1004 11:17:59.325626 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:17:59 crc kubenswrapper[4758]: E1004 11:17:59.327542 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:17:59 crc kubenswrapper[4758]: I1004 11:17:59.497009 4758 generic.go:334] "Generic (PLEG): container finished" podID="ceea8611-3691-4c29-b4ae-5c1e0d0bb396" containerID="ef246d3f0c5b1c6511eb2cf6c684b5de07bd7709425052e3cbeec76468a5cfee" exitCode=0 Oct 04 11:17:59 crc kubenswrapper[4758]: I1004 11:17:59.497068 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" event={"ID":"ceea8611-3691-4c29-b4ae-5c1e0d0bb396","Type":"ContainerDied","Data":"ef246d3f0c5b1c6511eb2cf6c684b5de07bd7709425052e3cbeec76468a5cfee"} Oct 04 11:18:00 crc kubenswrapper[4758]: I1004 11:18:00.032040 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-mgdlb"] Oct 04 11:18:00 crc kubenswrapper[4758]: I1004 11:18:00.041608 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-mgdlb"] Oct 04 11:18:00 crc kubenswrapper[4758]: I1004 11:18:00.952411 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.039086 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j2ws\" (UniqueName: \"kubernetes.io/projected/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-kube-api-access-7j2ws\") pod \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.039252 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-inventory\") pod \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.039313 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-ssh-key\") pod \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\" (UID: \"ceea8611-3691-4c29-b4ae-5c1e0d0bb396\") " Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.041498 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-h9jzh"] Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.054175 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kpx5z"] Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.067185 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-h9jzh"] Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.068548 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-kube-api-access-7j2ws" (OuterVolumeSpecName: "kube-api-access-7j2ws") pod "ceea8611-3691-4c29-b4ae-5c1e0d0bb396" (UID: "ceea8611-3691-4c29-b4ae-5c1e0d0bb396"). InnerVolumeSpecName "kube-api-access-7j2ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.081589 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kpx5z"] Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.082349 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ceea8611-3691-4c29-b4ae-5c1e0d0bb396" (UID: "ceea8611-3691-4c29-b4ae-5c1e0d0bb396"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.086366 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-inventory" (OuterVolumeSpecName: "inventory") pod "ceea8611-3691-4c29-b4ae-5c1e0d0bb396" (UID: "ceea8611-3691-4c29-b4ae-5c1e0d0bb396"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.141407 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j2ws\" (UniqueName: \"kubernetes.io/projected/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-kube-api-access-7j2ws\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.141448 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.141457 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ceea8611-3691-4c29-b4ae-5c1e0d0bb396-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.338124 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ddb1196-e3aa-4fe4-8f93-e29c25a37664" path="/var/lib/kubelet/pods/8ddb1196-e3aa-4fe4-8f93-e29c25a37664/volumes" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.343914 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c471b6cd-fc95-4ce7-847f-3ea9cc76d74a" path="/var/lib/kubelet/pods/c471b6cd-fc95-4ce7-847f-3ea9cc76d74a/volumes" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.347782 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca9af889-3123-43e5-b21b-c8cd9462563c" path="/var/lib/kubelet/pods/ca9af889-3123-43e5-b21b-c8cd9462563c/volumes" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.515718 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" event={"ID":"ceea8611-3691-4c29-b4ae-5c1e0d0bb396","Type":"ContainerDied","Data":"6e042a5a3057defece034649f77d2876984d11adbd3fba87452dbe74898621c2"} Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.515754 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e042a5a3057defece034649f77d2876984d11adbd3fba87452dbe74898621c2" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.515802 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-slrft" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.628305 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd"] Oct 04 11:18:01 crc kubenswrapper[4758]: E1004 11:18:01.628713 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceea8611-3691-4c29-b4ae-5c1e0d0bb396" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.628732 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceea8611-3691-4c29-b4ae-5c1e0d0bb396" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.628950 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceea8611-3691-4c29-b4ae-5c1e0d0bb396" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.629657 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.635710 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.635929 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.636057 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.636494 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.650694 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd"] Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.752376 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.752463 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvf9k\" (UniqueName: \"kubernetes.io/projected/5857921a-d32b-4089-98af-e1b14930e1e9-kube-api-access-vvf9k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.752512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.853755 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.853925 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.853999 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvf9k\" (UniqueName: \"kubernetes.io/projected/5857921a-d32b-4089-98af-e1b14930e1e9-kube-api-access-vvf9k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.857330 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.857472 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.882222 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvf9k\" (UniqueName: \"kubernetes.io/projected/5857921a-d32b-4089-98af-e1b14930e1e9-kube-api-access-vvf9k\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:01 crc kubenswrapper[4758]: I1004 11:18:01.944165 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:18:02 crc kubenswrapper[4758]: I1004 11:18:02.559698 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd"] Oct 04 11:18:03 crc kubenswrapper[4758]: I1004 11:18:03.533710 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" event={"ID":"5857921a-d32b-4089-98af-e1b14930e1e9","Type":"ContainerStarted","Data":"0db6ecc1ca20fe48bc7198bc0d65c41e65565017d2e7f4c7f7c03b68d69abfc3"} Oct 04 11:18:03 crc kubenswrapper[4758]: I1004 11:18:03.534059 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" event={"ID":"5857921a-d32b-4089-98af-e1b14930e1e9","Type":"ContainerStarted","Data":"abfab0a271c335f97187c4afd7108565888771e4e748d9f62e3821701cb5ca77"} Oct 04 11:18:03 crc kubenswrapper[4758]: I1004 11:18:03.555603 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" podStartSLOduration=2.111005368 podStartE2EDuration="2.555581493s" podCreationTimestamp="2025-10-04 11:18:01 +0000 UTC" firstStartedPulling="2025-10-04 11:18:02.5688551 +0000 UTC m=+1659.861505989" lastFinishedPulling="2025-10-04 11:18:03.013431205 +0000 UTC m=+1660.306082114" observedRunningTime="2025-10-04 11:18:03.547851867 +0000 UTC m=+1660.840502756" watchObservedRunningTime="2025-10-04 11:18:03.555581493 +0000 UTC m=+1660.848232382" Oct 04 11:18:11 crc kubenswrapper[4758]: I1004 11:18:11.326516 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:18:11 crc kubenswrapper[4758]: E1004 11:18:11.327270 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:18:21 crc kubenswrapper[4758]: I1004 11:18:21.048983 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-w9lbx"] Oct 04 11:18:21 crc kubenswrapper[4758]: I1004 11:18:21.056189 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-w9lbx"] Oct 04 11:18:21 crc kubenswrapper[4758]: I1004 11:18:21.342209 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d2a43a-542a-40c5-884f-e6540082151e" path="/var/lib/kubelet/pods/41d2a43a-542a-40c5-884f-e6540082151e/volumes" Oct 04 11:18:24 crc kubenswrapper[4758]: I1004 11:18:24.326426 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:18:24 crc kubenswrapper[4758]: E1004 11:18:24.327279 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.326919 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:18:35 crc kubenswrapper[4758]: E1004 11:18:35.327749 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.524815 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8hdtb"] Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.526871 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.535876 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hdtb"] Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.657480 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jztx8\" (UniqueName: \"kubernetes.io/projected/fdb218a3-1bd8-494b-9e02-e52e29771725-kube-api-access-jztx8\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.657572 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-utilities\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.657619 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-catalog-content\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.758939 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-utilities\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.759030 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-catalog-content\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.759178 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jztx8\" (UniqueName: \"kubernetes.io/projected/fdb218a3-1bd8-494b-9e02-e52e29771725-kube-api-access-jztx8\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.760150 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-utilities\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.760471 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-catalog-content\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.786902 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jztx8\" (UniqueName: \"kubernetes.io/projected/fdb218a3-1bd8-494b-9e02-e52e29771725-kube-api-access-jztx8\") pod \"redhat-marketplace-8hdtb\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:35 crc kubenswrapper[4758]: I1004 11:18:35.862335 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:36 crc kubenswrapper[4758]: I1004 11:18:36.386593 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hdtb"] Oct 04 11:18:36 crc kubenswrapper[4758]: I1004 11:18:36.832958 4758 generic.go:334] "Generic (PLEG): container finished" podID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerID="51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa" exitCode=0 Oct 04 11:18:36 crc kubenswrapper[4758]: I1004 11:18:36.833039 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hdtb" event={"ID":"fdb218a3-1bd8-494b-9e02-e52e29771725","Type":"ContainerDied","Data":"51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa"} Oct 04 11:18:36 crc kubenswrapper[4758]: I1004 11:18:36.833293 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hdtb" event={"ID":"fdb218a3-1bd8-494b-9e02-e52e29771725","Type":"ContainerStarted","Data":"80f1b54848beb1c4f546e4cb03c66f1c2291e6930330418455b84a318c453271"} Oct 04 11:18:38 crc kubenswrapper[4758]: I1004 11:18:38.851415 4758 generic.go:334] "Generic (PLEG): container finished" podID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerID="41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c" exitCode=0 Oct 04 11:18:38 crc kubenswrapper[4758]: I1004 11:18:38.851482 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hdtb" event={"ID":"fdb218a3-1bd8-494b-9e02-e52e29771725","Type":"ContainerDied","Data":"41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c"} Oct 04 11:18:39 crc kubenswrapper[4758]: I1004 11:18:39.861436 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hdtb" event={"ID":"fdb218a3-1bd8-494b-9e02-e52e29771725","Type":"ContainerStarted","Data":"310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02"} Oct 04 11:18:39 crc kubenswrapper[4758]: I1004 11:18:39.886283 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8hdtb" podStartSLOduration=2.044674529 podStartE2EDuration="4.886244s" podCreationTimestamp="2025-10-04 11:18:35 +0000 UTC" firstStartedPulling="2025-10-04 11:18:36.834445396 +0000 UTC m=+1694.127096295" lastFinishedPulling="2025-10-04 11:18:39.676014877 +0000 UTC m=+1696.968665766" observedRunningTime="2025-10-04 11:18:39.881999167 +0000 UTC m=+1697.174650056" watchObservedRunningTime="2025-10-04 11:18:39.886244 +0000 UTC m=+1697.178894889" Oct 04 11:18:45 crc kubenswrapper[4758]: I1004 11:18:45.863053 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:45 crc kubenswrapper[4758]: I1004 11:18:45.863728 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:45 crc kubenswrapper[4758]: I1004 11:18:45.913571 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:45 crc kubenswrapper[4758]: I1004 11:18:45.977185 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:46 crc kubenswrapper[4758]: I1004 11:18:46.154477 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hdtb"] Oct 04 11:18:47 crc kubenswrapper[4758]: I1004 11:18:47.947219 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8hdtb" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="registry-server" containerID="cri-o://310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02" gracePeriod=2 Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.442769 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.598940 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-catalog-content\") pod \"fdb218a3-1bd8-494b-9e02-e52e29771725\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.599077 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jztx8\" (UniqueName: \"kubernetes.io/projected/fdb218a3-1bd8-494b-9e02-e52e29771725-kube-api-access-jztx8\") pod \"fdb218a3-1bd8-494b-9e02-e52e29771725\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.599271 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-utilities\") pod \"fdb218a3-1bd8-494b-9e02-e52e29771725\" (UID: \"fdb218a3-1bd8-494b-9e02-e52e29771725\") " Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.600388 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-utilities" (OuterVolumeSpecName: "utilities") pod "fdb218a3-1bd8-494b-9e02-e52e29771725" (UID: "fdb218a3-1bd8-494b-9e02-e52e29771725"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.613331 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb218a3-1bd8-494b-9e02-e52e29771725-kube-api-access-jztx8" (OuterVolumeSpecName: "kube-api-access-jztx8") pod "fdb218a3-1bd8-494b-9e02-e52e29771725" (UID: "fdb218a3-1bd8-494b-9e02-e52e29771725"). InnerVolumeSpecName "kube-api-access-jztx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.618161 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdb218a3-1bd8-494b-9e02-e52e29771725" (UID: "fdb218a3-1bd8-494b-9e02-e52e29771725"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.702122 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jztx8\" (UniqueName: \"kubernetes.io/projected/fdb218a3-1bd8-494b-9e02-e52e29771725-kube-api-access-jztx8\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.702177 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.702191 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdb218a3-1bd8-494b-9e02-e52e29771725-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.958149 4758 generic.go:334] "Generic (PLEG): container finished" podID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerID="310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02" exitCode=0 Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.958356 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hdtb" event={"ID":"fdb218a3-1bd8-494b-9e02-e52e29771725","Type":"ContainerDied","Data":"310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02"} Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.958509 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8hdtb" event={"ID":"fdb218a3-1bd8-494b-9e02-e52e29771725","Type":"ContainerDied","Data":"80f1b54848beb1c4f546e4cb03c66f1c2291e6930330418455b84a318c453271"} Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.958544 4758 scope.go:117] "RemoveContainer" containerID="310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.960199 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8hdtb" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.987460 4758 scope.go:117] "RemoveContainer" containerID="41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c" Oct 04 11:18:48 crc kubenswrapper[4758]: I1004 11:18:48.999419 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hdtb"] Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.015149 4758 scope.go:117] "RemoveContainer" containerID="51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.015586 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8hdtb"] Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.077805 4758 scope.go:117] "RemoveContainer" containerID="310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02" Oct 04 11:18:49 crc kubenswrapper[4758]: E1004 11:18:49.078366 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02\": container with ID starting with 310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02 not found: ID does not exist" containerID="310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.078410 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02"} err="failed to get container status \"310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02\": rpc error: code = NotFound desc = could not find container \"310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02\": container with ID starting with 310c31954e7a44f01ca1c801c8d3c28f99d1e5599dca3d9495d1ca0586fc7d02 not found: ID does not exist" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.078437 4758 scope.go:117] "RemoveContainer" containerID="41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c" Oct 04 11:18:49 crc kubenswrapper[4758]: E1004 11:18:49.078889 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c\": container with ID starting with 41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c not found: ID does not exist" containerID="41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.079032 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c"} err="failed to get container status \"41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c\": rpc error: code = NotFound desc = could not find container \"41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c\": container with ID starting with 41d313c3aa81ecbb285c55a669e47e668c17e7938b77e9f6db6be19f58c7dc2c not found: ID does not exist" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.079177 4758 scope.go:117] "RemoveContainer" containerID="51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa" Oct 04 11:18:49 crc kubenswrapper[4758]: E1004 11:18:49.079636 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa\": container with ID starting with 51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa not found: ID does not exist" containerID="51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.079734 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa"} err="failed to get container status \"51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa\": rpc error: code = NotFound desc = could not find container \"51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa\": container with ID starting with 51639236ca2b9e702d18f109d1f37e32a99a342fafe1a3ea071486ed4866bcfa not found: ID does not exist" Oct 04 11:18:49 crc kubenswrapper[4758]: I1004 11:18:49.338508 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" path="/var/lib/kubelet/pods/fdb218a3-1bd8-494b-9e02-e52e29771725/volumes" Oct 04 11:18:50 crc kubenswrapper[4758]: I1004 11:18:50.326669 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:18:50 crc kubenswrapper[4758]: E1004 11:18:50.327000 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.042918 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hw2fc"] Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.051434 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-8tnvw"] Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.060842 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hw2fc"] Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.071592 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-8tnvw"] Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.084320 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-s9zzx"] Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.089391 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-s9zzx"] Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.340563 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53b6351b-2066-44a5-90fe-f06e671b9452" path="/var/lib/kubelet/pods/53b6351b-2066-44a5-90fe-f06e671b9452/volumes" Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.344373 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6" path="/var/lib/kubelet/pods/6fc5fb58-7cb5-4ec2-aa41-84ef5a3eddb6/volumes" Oct 04 11:18:55 crc kubenswrapper[4758]: I1004 11:18:55.360237 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f9b2b3-f0ed-4fb8-837f-421ebaf92579" path="/var/lib/kubelet/pods/b6f9b2b3-f0ed-4fb8-837f-421ebaf92579/volumes" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.070909 4758 scope.go:117] "RemoveContainer" containerID="b95adcbdc5be222cb3cb312af727b67de1985c050a2a53a7c6d7552caccd835c" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.123464 4758 scope.go:117] "RemoveContainer" containerID="583899d86731878285b20ae19e2cfa6e62a111a204b272d7cbfe8499fb447beb" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.190292 4758 scope.go:117] "RemoveContainer" containerID="13263f9591ec9b17b5cce4b8a589f37bb71f6bd905aec33d3881ef7c493d85b6" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.221550 4758 scope.go:117] "RemoveContainer" containerID="ad746931127bfee5948e7431d4b3fbb488edcc36ffbecaf167f89614124c4f3b" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.267042 4758 scope.go:117] "RemoveContainer" containerID="e521a77eb4ba65d1251271bec3d9d0e60da5c75272ab719171bf4e32e97ec946" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.317430 4758 scope.go:117] "RemoveContainer" containerID="d86b51fa03ce7d44dda6d844d4981bf2c956086380b7941878bfb73fe9654feb" Oct 04 11:18:57 crc kubenswrapper[4758]: I1004 11:18:57.357236 4758 scope.go:117] "RemoveContainer" containerID="75a4965a9e350f492a8524fe9c228a230ff9b1601d74768f07ac8063c6626ac6" Oct 04 11:19:02 crc kubenswrapper[4758]: I1004 11:19:02.325705 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:19:02 crc kubenswrapper[4758]: E1004 11:19:02.326703 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:19:11 crc kubenswrapper[4758]: I1004 11:19:11.042764 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-1ee5-account-create-zfbg7"] Oct 04 11:19:11 crc kubenswrapper[4758]: I1004 11:19:11.053179 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-18f9-account-create-46nw9"] Oct 04 11:19:11 crc kubenswrapper[4758]: I1004 11:19:11.078973 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-1ee5-account-create-zfbg7"] Oct 04 11:19:11 crc kubenswrapper[4758]: I1004 11:19:11.079043 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-18f9-account-create-46nw9"] Oct 04 11:19:11 crc kubenswrapper[4758]: I1004 11:19:11.336025 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a17d34-49c8-4ad5-8f2d-86b9b665dfa2" path="/var/lib/kubelet/pods/51a17d34-49c8-4ad5-8f2d-86b9b665dfa2/volumes" Oct 04 11:19:11 crc kubenswrapper[4758]: I1004 11:19:11.338356 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd746b6d-78c5-4ef6-a466-ad7ed7d346cf" path="/var/lib/kubelet/pods/cd746b6d-78c5-4ef6-a466-ad7ed7d346cf/volumes" Oct 04 11:19:12 crc kubenswrapper[4758]: I1004 11:19:12.027476 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-94f8-account-create-7474t"] Oct 04 11:19:12 crc kubenswrapper[4758]: I1004 11:19:12.041319 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-94f8-account-create-7474t"] Oct 04 11:19:13 crc kubenswrapper[4758]: I1004 11:19:13.337256 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f503482b-df1b-4849-ad88-7a723462b9d4" path="/var/lib/kubelet/pods/f503482b-df1b-4849-ad88-7a723462b9d4/volumes" Oct 04 11:19:17 crc kubenswrapper[4758]: I1004 11:19:17.327251 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:19:17 crc kubenswrapper[4758]: E1004 11:19:17.327960 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:19:19 crc kubenswrapper[4758]: I1004 11:19:19.247412 4758 generic.go:334] "Generic (PLEG): container finished" podID="5857921a-d32b-4089-98af-e1b14930e1e9" containerID="0db6ecc1ca20fe48bc7198bc0d65c41e65565017d2e7f4c7f7c03b68d69abfc3" exitCode=0 Oct 04 11:19:19 crc kubenswrapper[4758]: I1004 11:19:19.247536 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" event={"ID":"5857921a-d32b-4089-98af-e1b14930e1e9","Type":"ContainerDied","Data":"0db6ecc1ca20fe48bc7198bc0d65c41e65565017d2e7f4c7f7c03b68d69abfc3"} Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.673024 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.815571 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvf9k\" (UniqueName: \"kubernetes.io/projected/5857921a-d32b-4089-98af-e1b14930e1e9-kube-api-access-vvf9k\") pod \"5857921a-d32b-4089-98af-e1b14930e1e9\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.815822 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-inventory\") pod \"5857921a-d32b-4089-98af-e1b14930e1e9\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.815859 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-ssh-key\") pod \"5857921a-d32b-4089-98af-e1b14930e1e9\" (UID: \"5857921a-d32b-4089-98af-e1b14930e1e9\") " Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.822290 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5857921a-d32b-4089-98af-e1b14930e1e9-kube-api-access-vvf9k" (OuterVolumeSpecName: "kube-api-access-vvf9k") pod "5857921a-d32b-4089-98af-e1b14930e1e9" (UID: "5857921a-d32b-4089-98af-e1b14930e1e9"). InnerVolumeSpecName "kube-api-access-vvf9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.848204 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-inventory" (OuterVolumeSpecName: "inventory") pod "5857921a-d32b-4089-98af-e1b14930e1e9" (UID: "5857921a-d32b-4089-98af-e1b14930e1e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.867370 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5857921a-d32b-4089-98af-e1b14930e1e9" (UID: "5857921a-d32b-4089-98af-e1b14930e1e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.917613 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.917838 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5857921a-d32b-4089-98af-e1b14930e1e9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:20 crc kubenswrapper[4758]: I1004 11:19:20.917904 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvf9k\" (UniqueName: \"kubernetes.io/projected/5857921a-d32b-4089-98af-e1b14930e1e9-kube-api-access-vvf9k\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.265275 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" event={"ID":"5857921a-d32b-4089-98af-e1b14930e1e9","Type":"ContainerDied","Data":"abfab0a271c335f97187c4afd7108565888771e4e748d9f62e3821701cb5ca77"} Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.265646 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abfab0a271c335f97187c4afd7108565888771e4e748d9f62e3821701cb5ca77" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.265342 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.376858 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh"] Oct 04 11:19:21 crc kubenswrapper[4758]: E1004 11:19:21.377336 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="extract-utilities" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.377358 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="extract-utilities" Oct 04 11:19:21 crc kubenswrapper[4758]: E1004 11:19:21.377387 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5857921a-d32b-4089-98af-e1b14930e1e9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.377399 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="5857921a-d32b-4089-98af-e1b14930e1e9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:19:21 crc kubenswrapper[4758]: E1004 11:19:21.377415 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="extract-content" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.377424 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="extract-content" Oct 04 11:19:21 crc kubenswrapper[4758]: E1004 11:19:21.377440 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="registry-server" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.377447 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="registry-server" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.377692 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb218a3-1bd8-494b-9e02-e52e29771725" containerName="registry-server" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.377711 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="5857921a-d32b-4089-98af-e1b14930e1e9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.378844 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.383636 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.383850 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.385405 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.385444 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.391227 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh"] Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.426828 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.427217 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k65pp\" (UniqueName: \"kubernetes.io/projected/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-kube-api-access-k65pp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.427275 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.528589 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k65pp\" (UniqueName: \"kubernetes.io/projected/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-kube-api-access-k65pp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.528677 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.528906 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.533526 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.533746 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.550687 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k65pp\" (UniqueName: \"kubernetes.io/projected/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-kube-api-access-k65pp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:21 crc kubenswrapper[4758]: I1004 11:19:21.698598 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:22 crc kubenswrapper[4758]: I1004 11:19:22.256598 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh"] Oct 04 11:19:22 crc kubenswrapper[4758]: I1004 11:19:22.275698 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" event={"ID":"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1","Type":"ContainerStarted","Data":"b855c796a6db40e6e5a4f4583618fbce9042ba2da65faae7cad7ff2645f051f2"} Oct 04 11:19:24 crc kubenswrapper[4758]: I1004 11:19:24.301469 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" event={"ID":"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1","Type":"ContainerStarted","Data":"4e9be7f091d2855d930a3e2a2ad201f510445996c65fd414f11db7227d98c751"} Oct 04 11:19:24 crc kubenswrapper[4758]: I1004 11:19:24.322664 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" podStartSLOduration=2.466381442 podStartE2EDuration="3.322648191s" podCreationTimestamp="2025-10-04 11:19:21 +0000 UTC" firstStartedPulling="2025-10-04 11:19:22.26092472 +0000 UTC m=+1739.553575609" lastFinishedPulling="2025-10-04 11:19:23.117191469 +0000 UTC m=+1740.409842358" observedRunningTime="2025-10-04 11:19:24.319874868 +0000 UTC m=+1741.612525757" watchObservedRunningTime="2025-10-04 11:19:24.322648191 +0000 UTC m=+1741.615299080" Oct 04 11:19:29 crc kubenswrapper[4758]: I1004 11:19:29.348304 4758 generic.go:334] "Generic (PLEG): container finished" podID="a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" containerID="4e9be7f091d2855d930a3e2a2ad201f510445996c65fd414f11db7227d98c751" exitCode=0 Oct 04 11:19:29 crc kubenswrapper[4758]: I1004 11:19:29.348907 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" event={"ID":"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1","Type":"ContainerDied","Data":"4e9be7f091d2855d930a3e2a2ad201f510445996c65fd414f11db7227d98c751"} Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.782472 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.911494 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-inventory\") pod \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.911556 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k65pp\" (UniqueName: \"kubernetes.io/projected/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-kube-api-access-k65pp\") pod \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.911670 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-ssh-key\") pod \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\" (UID: \"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1\") " Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.930185 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-kube-api-access-k65pp" (OuterVolumeSpecName: "kube-api-access-k65pp") pod "a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" (UID: "a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1"). InnerVolumeSpecName "kube-api-access-k65pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.938465 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-inventory" (OuterVolumeSpecName: "inventory") pod "a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" (UID: "a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:30 crc kubenswrapper[4758]: I1004 11:19:30.938783 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" (UID: "a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.013291 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.013340 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k65pp\" (UniqueName: \"kubernetes.io/projected/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-kube-api-access-k65pp\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.013352 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.325579 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:19:31 crc kubenswrapper[4758]: E1004 11:19:31.325916 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.366037 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" event={"ID":"a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1","Type":"ContainerDied","Data":"b855c796a6db40e6e5a4f4583618fbce9042ba2da65faae7cad7ff2645f051f2"} Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.366080 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b855c796a6db40e6e5a4f4583618fbce9042ba2da65faae7cad7ff2645f051f2" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.366088 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.456169 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q"] Oct 04 11:19:31 crc kubenswrapper[4758]: E1004 11:19:31.456606 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.456624 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.456842 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.457530 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.463642 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.463813 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.464076 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.473131 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.480743 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q"] Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.522972 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.523132 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.523197 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbvsz\" (UniqueName: \"kubernetes.io/projected/89420680-fecc-433a-a18e-5f2b65b98dfe-kube-api-access-xbvsz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.624886 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.624993 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbvsz\" (UniqueName: \"kubernetes.io/projected/89420680-fecc-433a-a18e-5f2b65b98dfe-kube-api-access-xbvsz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.625041 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.630211 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.634603 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.645592 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbvsz\" (UniqueName: \"kubernetes.io/projected/89420680-fecc-433a-a18e-5f2b65b98dfe-kube-api-access-xbvsz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mlz4q\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:31 crc kubenswrapper[4758]: I1004 11:19:31.786779 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:19:32 crc kubenswrapper[4758]: I1004 11:19:32.341025 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q"] Oct 04 11:19:32 crc kubenswrapper[4758]: I1004 11:19:32.376790 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" event={"ID":"89420680-fecc-433a-a18e-5f2b65b98dfe","Type":"ContainerStarted","Data":"baf0205140895622e8015268bae4c445948d04294ce6e31f9c73492b74ec7675"} Oct 04 11:19:33 crc kubenswrapper[4758]: I1004 11:19:33.384955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" event={"ID":"89420680-fecc-433a-a18e-5f2b65b98dfe","Type":"ContainerStarted","Data":"411cb9a1cc6b047dff9f9c82ebf6c0e23d90b9cd42ce9a518ce1b41d94189a63"} Oct 04 11:19:33 crc kubenswrapper[4758]: I1004 11:19:33.428202 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" podStartSLOduration=1.92630835 podStartE2EDuration="2.428187008s" podCreationTimestamp="2025-10-04 11:19:31 +0000 UTC" firstStartedPulling="2025-10-04 11:19:32.356098027 +0000 UTC m=+1749.648748916" lastFinishedPulling="2025-10-04 11:19:32.857976675 +0000 UTC m=+1750.150627574" observedRunningTime="2025-10-04 11:19:33.426502823 +0000 UTC m=+1750.719153712" watchObservedRunningTime="2025-10-04 11:19:33.428187008 +0000 UTC m=+1750.720837897" Oct 04 11:19:42 crc kubenswrapper[4758]: I1004 11:19:42.096603 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8tq5p"] Oct 04 11:19:42 crc kubenswrapper[4758]: I1004 11:19:42.143068 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8tq5p"] Oct 04 11:19:43 crc kubenswrapper[4758]: I1004 11:19:43.331245 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:19:43 crc kubenswrapper[4758]: E1004 11:19:43.331764 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:19:43 crc kubenswrapper[4758]: I1004 11:19:43.371138 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56103956-1715-4a93-9784-70dd825c43a2" path="/var/lib/kubelet/pods/56103956-1715-4a93-9784-70dd825c43a2/volumes" Oct 04 11:19:56 crc kubenswrapper[4758]: I1004 11:19:56.326503 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:19:56 crc kubenswrapper[4758]: E1004 11:19:56.327431 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:19:57 crc kubenswrapper[4758]: I1004 11:19:57.524926 4758 scope.go:117] "RemoveContainer" containerID="d9a6f7fe3bc5b63061bb8c05a6529699471f771b3f5063ad38e7d9af73ebd0f3" Oct 04 11:19:57 crc kubenswrapper[4758]: I1004 11:19:57.563418 4758 scope.go:117] "RemoveContainer" containerID="ffe8fedcfe68f06d92fb98d31f6c1b7f90a7ccb82d7b5662cd6ce39b8cd39fc2" Oct 04 11:19:57 crc kubenswrapper[4758]: I1004 11:19:57.633374 4758 scope.go:117] "RemoveContainer" containerID="fc928abe78917a701c13580ece3f66592a29df27fe9f44b82ae173b257641ef9" Oct 04 11:19:57 crc kubenswrapper[4758]: I1004 11:19:57.745251 4758 scope.go:117] "RemoveContainer" containerID="9ffc9eaa9ef2818852dcbdd662ca88e1126b420be8d0d631c7afdc3d45f32d27" Oct 04 11:20:02 crc kubenswrapper[4758]: I1004 11:20:02.055156 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-4p4g9"] Oct 04 11:20:02 crc kubenswrapper[4758]: I1004 11:20:02.067288 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-4p4g9"] Oct 04 11:20:03 crc kubenswrapper[4758]: I1004 11:20:03.032700 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gn24j"] Oct 04 11:20:03 crc kubenswrapper[4758]: I1004 11:20:03.043722 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gn24j"] Oct 04 11:20:03 crc kubenswrapper[4758]: I1004 11:20:03.337405 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702dad08-3e1d-4118-9ecd-6c1882b8188d" path="/var/lib/kubelet/pods/702dad08-3e1d-4118-9ecd-6c1882b8188d/volumes" Oct 04 11:20:03 crc kubenswrapper[4758]: I1004 11:20:03.356663 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac470595-559d-4607-9113-450378bcaf13" path="/var/lib/kubelet/pods/ac470595-559d-4607-9113-450378bcaf13/volumes" Oct 04 11:20:08 crc kubenswrapper[4758]: I1004 11:20:08.326612 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:20:08 crc kubenswrapper[4758]: E1004 11:20:08.327566 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:20:15 crc kubenswrapper[4758]: I1004 11:20:15.734349 4758 generic.go:334] "Generic (PLEG): container finished" podID="89420680-fecc-433a-a18e-5f2b65b98dfe" containerID="411cb9a1cc6b047dff9f9c82ebf6c0e23d90b9cd42ce9a518ce1b41d94189a63" exitCode=0 Oct 04 11:20:15 crc kubenswrapper[4758]: I1004 11:20:15.734449 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" event={"ID":"89420680-fecc-433a-a18e-5f2b65b98dfe","Type":"ContainerDied","Data":"411cb9a1cc6b047dff9f9c82ebf6c0e23d90b9cd42ce9a518ce1b41d94189a63"} Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.136422 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.196837 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-inventory\") pod \"89420680-fecc-433a-a18e-5f2b65b98dfe\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.196912 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-ssh-key\") pod \"89420680-fecc-433a-a18e-5f2b65b98dfe\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.197151 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbvsz\" (UniqueName: \"kubernetes.io/projected/89420680-fecc-433a-a18e-5f2b65b98dfe-kube-api-access-xbvsz\") pod \"89420680-fecc-433a-a18e-5f2b65b98dfe\" (UID: \"89420680-fecc-433a-a18e-5f2b65b98dfe\") " Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.205420 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89420680-fecc-433a-a18e-5f2b65b98dfe-kube-api-access-xbvsz" (OuterVolumeSpecName: "kube-api-access-xbvsz") pod "89420680-fecc-433a-a18e-5f2b65b98dfe" (UID: "89420680-fecc-433a-a18e-5f2b65b98dfe"). InnerVolumeSpecName "kube-api-access-xbvsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.229056 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "89420680-fecc-433a-a18e-5f2b65b98dfe" (UID: "89420680-fecc-433a-a18e-5f2b65b98dfe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.232956 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-inventory" (OuterVolumeSpecName: "inventory") pod "89420680-fecc-433a-a18e-5f2b65b98dfe" (UID: "89420680-fecc-433a-a18e-5f2b65b98dfe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.300179 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbvsz\" (UniqueName: \"kubernetes.io/projected/89420680-fecc-433a-a18e-5f2b65b98dfe-kube-api-access-xbvsz\") on node \"crc\" DevicePath \"\"" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.300258 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.300269 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89420680-fecc-433a-a18e-5f2b65b98dfe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.772989 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" event={"ID":"89420680-fecc-433a-a18e-5f2b65b98dfe","Type":"ContainerDied","Data":"baf0205140895622e8015268bae4c445948d04294ce6e31f9c73492b74ec7675"} Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.773350 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baf0205140895622e8015268bae4c445948d04294ce6e31f9c73492b74ec7675" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.773415 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mlz4q" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.853722 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt"] Oct 04 11:20:17 crc kubenswrapper[4758]: E1004 11:20:17.854569 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89420680-fecc-433a-a18e-5f2b65b98dfe" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.854589 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="89420680-fecc-433a-a18e-5f2b65b98dfe" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.854833 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="89420680-fecc-433a-a18e-5f2b65b98dfe" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.855788 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.864027 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt"] Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.882320 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.882403 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.882536 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.882833 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.915626 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.915729 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-786fp\" (UniqueName: \"kubernetes.io/projected/e9547006-e0e0-4294-a079-7bb90e5e44ae-kube-api-access-786fp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:17 crc kubenswrapper[4758]: I1004 11:20:17.915781 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.017361 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.017427 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-786fp\" (UniqueName: \"kubernetes.io/projected/e9547006-e0e0-4294-a079-7bb90e5e44ae-kube-api-access-786fp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.017476 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.022533 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.022638 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.034077 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-786fp\" (UniqueName: \"kubernetes.io/projected/e9547006-e0e0-4294-a079-7bb90e5e44ae-kube-api-access-786fp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:18 crc kubenswrapper[4758]: I1004 11:20:18.200444 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:20:19 crc kubenswrapper[4758]: I1004 11:20:18.743027 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt"] Oct 04 11:20:19 crc kubenswrapper[4758]: W1004 11:20:18.751346 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9547006_e0e0_4294_a079_7bb90e5e44ae.slice/crio-6d232f18a77ebbad9be4f7d891e46e92b2aa39e50e5bcd8baeb6dc2551192286 WatchSource:0}: Error finding container 6d232f18a77ebbad9be4f7d891e46e92b2aa39e50e5bcd8baeb6dc2551192286: Status 404 returned error can't find the container with id 6d232f18a77ebbad9be4f7d891e46e92b2aa39e50e5bcd8baeb6dc2551192286 Oct 04 11:20:19 crc kubenswrapper[4758]: I1004 11:20:18.800176 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" event={"ID":"e9547006-e0e0-4294-a079-7bb90e5e44ae","Type":"ContainerStarted","Data":"6d232f18a77ebbad9be4f7d891e46e92b2aa39e50e5bcd8baeb6dc2551192286"} Oct 04 11:20:19 crc kubenswrapper[4758]: I1004 11:20:19.812290 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" event={"ID":"e9547006-e0e0-4294-a079-7bb90e5e44ae","Type":"ContainerStarted","Data":"304525401a27600a3786f84ab2b1de3bdb088ecee6e7007da0043e6dcd010f0f"} Oct 04 11:20:19 crc kubenswrapper[4758]: I1004 11:20:19.837213 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" podStartSLOduration=2.199520397 podStartE2EDuration="2.837197781s" podCreationTimestamp="2025-10-04 11:20:17 +0000 UTC" firstStartedPulling="2025-10-04 11:20:18.753217815 +0000 UTC m=+1796.045868704" lastFinishedPulling="2025-10-04 11:20:19.390895179 +0000 UTC m=+1796.683546088" observedRunningTime="2025-10-04 11:20:19.835891517 +0000 UTC m=+1797.128542436" watchObservedRunningTime="2025-10-04 11:20:19.837197781 +0000 UTC m=+1797.129848670" Oct 04 11:20:22 crc kubenswrapper[4758]: I1004 11:20:22.326749 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:20:22 crc kubenswrapper[4758]: E1004 11:20:22.327494 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:20:34 crc kubenswrapper[4758]: I1004 11:20:34.325133 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:20:34 crc kubenswrapper[4758]: E1004 11:20:34.325840 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:20:46 crc kubenswrapper[4758]: I1004 11:20:46.046896 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-kw82l"] Oct 04 11:20:46 crc kubenswrapper[4758]: I1004 11:20:46.056609 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-kw82l"] Oct 04 11:20:47 crc kubenswrapper[4758]: I1004 11:20:47.341242 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c16e6f6-a6f4-49e4-b8e7-df0d923e801b" path="/var/lib/kubelet/pods/3c16e6f6-a6f4-49e4-b8e7-df0d923e801b/volumes" Oct 04 11:20:48 crc kubenswrapper[4758]: I1004 11:20:48.325845 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:20:48 crc kubenswrapper[4758]: E1004 11:20:48.326445 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:20:57 crc kubenswrapper[4758]: I1004 11:20:57.939871 4758 scope.go:117] "RemoveContainer" containerID="7ccc35e0e3932fe89919b7a550264f5d3ffe2e6a8bf754440f7b9e225b157765" Oct 04 11:20:58 crc kubenswrapper[4758]: I1004 11:20:58.008610 4758 scope.go:117] "RemoveContainer" containerID="08bc7f34c9986fcd62327c5e0164d0c73c636de627d599923e049bf336c847f5" Oct 04 11:20:58 crc kubenswrapper[4758]: I1004 11:20:58.088929 4758 scope.go:117] "RemoveContainer" containerID="b74dbf4be71a1fba3808e336d64557bae7afcd821ff80c44aad9c60a32d5769f" Oct 04 11:21:01 crc kubenswrapper[4758]: I1004 11:21:01.326906 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:21:01 crc kubenswrapper[4758]: E1004 11:21:01.327523 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:21:14 crc kubenswrapper[4758]: I1004 11:21:14.326762 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:21:14 crc kubenswrapper[4758]: E1004 11:21:14.327941 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:21:19 crc kubenswrapper[4758]: I1004 11:21:19.358083 4758 generic.go:334] "Generic (PLEG): container finished" podID="e9547006-e0e0-4294-a079-7bb90e5e44ae" containerID="304525401a27600a3786f84ab2b1de3bdb088ecee6e7007da0043e6dcd010f0f" exitCode=2 Oct 04 11:21:19 crc kubenswrapper[4758]: I1004 11:21:19.358186 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" event={"ID":"e9547006-e0e0-4294-a079-7bb90e5e44ae","Type":"ContainerDied","Data":"304525401a27600a3786f84ab2b1de3bdb088ecee6e7007da0043e6dcd010f0f"} Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.780031 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.861558 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-inventory\") pod \"e9547006-e0e0-4294-a079-7bb90e5e44ae\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.861758 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-786fp\" (UniqueName: \"kubernetes.io/projected/e9547006-e0e0-4294-a079-7bb90e5e44ae-kube-api-access-786fp\") pod \"e9547006-e0e0-4294-a079-7bb90e5e44ae\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.861822 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-ssh-key\") pod \"e9547006-e0e0-4294-a079-7bb90e5e44ae\" (UID: \"e9547006-e0e0-4294-a079-7bb90e5e44ae\") " Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.868148 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9547006-e0e0-4294-a079-7bb90e5e44ae-kube-api-access-786fp" (OuterVolumeSpecName: "kube-api-access-786fp") pod "e9547006-e0e0-4294-a079-7bb90e5e44ae" (UID: "e9547006-e0e0-4294-a079-7bb90e5e44ae"). InnerVolumeSpecName "kube-api-access-786fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.894329 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-inventory" (OuterVolumeSpecName: "inventory") pod "e9547006-e0e0-4294-a079-7bb90e5e44ae" (UID: "e9547006-e0e0-4294-a079-7bb90e5e44ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.895858 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e9547006-e0e0-4294-a079-7bb90e5e44ae" (UID: "e9547006-e0e0-4294-a079-7bb90e5e44ae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.964616 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.964681 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9547006-e0e0-4294-a079-7bb90e5e44ae-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:21:20 crc kubenswrapper[4758]: I1004 11:21:20.964702 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-786fp\" (UniqueName: \"kubernetes.io/projected/e9547006-e0e0-4294-a079-7bb90e5e44ae-kube-api-access-786fp\") on node \"crc\" DevicePath \"\"" Oct 04 11:21:21 crc kubenswrapper[4758]: I1004 11:21:21.379146 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" event={"ID":"e9547006-e0e0-4294-a079-7bb90e5e44ae","Type":"ContainerDied","Data":"6d232f18a77ebbad9be4f7d891e46e92b2aa39e50e5bcd8baeb6dc2551192286"} Oct 04 11:21:21 crc kubenswrapper[4758]: I1004 11:21:21.379219 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d232f18a77ebbad9be4f7d891e46e92b2aa39e50e5bcd8baeb6dc2551192286" Oct 04 11:21:21 crc kubenswrapper[4758]: I1004 11:21:21.379192 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt" Oct 04 11:21:26 crc kubenswrapper[4758]: I1004 11:21:26.326641 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:21:26 crc kubenswrapper[4758]: E1004 11:21:26.327671 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.046074 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv"] Oct 04 11:21:28 crc kubenswrapper[4758]: E1004 11:21:28.047046 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9547006-e0e0-4294-a079-7bb90e5e44ae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.047070 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9547006-e0e0-4294-a079-7bb90e5e44ae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.047499 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9547006-e0e0-4294-a079-7bb90e5e44ae" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.048565 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.068475 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv"] Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.088775 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.089309 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.089463 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.089614 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.205649 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b96j7\" (UniqueName: \"kubernetes.io/projected/c9446f9a-a59b-485b-be69-82419c46d6cd-kube-api-access-b96j7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.205726 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.205893 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.307398 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.307615 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.307750 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b96j7\" (UniqueName: \"kubernetes.io/projected/c9446f9a-a59b-485b-be69-82419c46d6cd-kube-api-access-b96j7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.319659 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.323430 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.325214 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b96j7\" (UniqueName: \"kubernetes.io/projected/c9446f9a-a59b-485b-be69-82419c46d6cd-kube-api-access-b96j7\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.420423 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.967530 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv"] Oct 04 11:21:28 crc kubenswrapper[4758]: I1004 11:21:28.976402 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:21:29 crc kubenswrapper[4758]: I1004 11:21:29.453833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" event={"ID":"c9446f9a-a59b-485b-be69-82419c46d6cd","Type":"ContainerStarted","Data":"0ec832cbda89787949524b5e688d9788d09f570486d9684b19b0c2bd309a2c92"} Oct 04 11:21:30 crc kubenswrapper[4758]: I1004 11:21:30.465258 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" event={"ID":"c9446f9a-a59b-485b-be69-82419c46d6cd","Type":"ContainerStarted","Data":"f129241ebea9cee06e85279866aa67b6131c56481015316ad4df4e5105709ee6"} Oct 04 11:21:30 crc kubenswrapper[4758]: I1004 11:21:30.499321 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" podStartSLOduration=1.990601171 podStartE2EDuration="2.499212348s" podCreationTimestamp="2025-10-04 11:21:28 +0000 UTC" firstStartedPulling="2025-10-04 11:21:28.976203793 +0000 UTC m=+1866.268854682" lastFinishedPulling="2025-10-04 11:21:29.48481493 +0000 UTC m=+1866.777465859" observedRunningTime="2025-10-04 11:21:30.490357823 +0000 UTC m=+1867.783008742" watchObservedRunningTime="2025-10-04 11:21:30.499212348 +0000 UTC m=+1867.791863237" Oct 04 11:21:38 crc kubenswrapper[4758]: I1004 11:21:38.327067 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:21:38 crc kubenswrapper[4758]: E1004 11:21:38.327827 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:21:51 crc kubenswrapper[4758]: I1004 11:21:51.326353 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:21:51 crc kubenswrapper[4758]: E1004 11:21:51.327446 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:22:03 crc kubenswrapper[4758]: I1004 11:22:03.332497 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:22:03 crc kubenswrapper[4758]: I1004 11:22:03.761064 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"8fb28a3d28b026315f63e9ed9f93326a8b51ed6380ad4d69c7930ab2c772926d"} Oct 04 11:22:22 crc kubenswrapper[4758]: I1004 11:22:22.931136 4758 generic.go:334] "Generic (PLEG): container finished" podID="c9446f9a-a59b-485b-be69-82419c46d6cd" containerID="f129241ebea9cee06e85279866aa67b6131c56481015316ad4df4e5105709ee6" exitCode=0 Oct 04 11:22:22 crc kubenswrapper[4758]: I1004 11:22:22.931143 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" event={"ID":"c9446f9a-a59b-485b-be69-82419c46d6cd","Type":"ContainerDied","Data":"f129241ebea9cee06e85279866aa67b6131c56481015316ad4df4e5105709ee6"} Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.386978 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.553510 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-ssh-key\") pod \"c9446f9a-a59b-485b-be69-82419c46d6cd\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.553612 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b96j7\" (UniqueName: \"kubernetes.io/projected/c9446f9a-a59b-485b-be69-82419c46d6cd-kube-api-access-b96j7\") pod \"c9446f9a-a59b-485b-be69-82419c46d6cd\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.553673 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-inventory\") pod \"c9446f9a-a59b-485b-be69-82419c46d6cd\" (UID: \"c9446f9a-a59b-485b-be69-82419c46d6cd\") " Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.586338 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9446f9a-a59b-485b-be69-82419c46d6cd-kube-api-access-b96j7" (OuterVolumeSpecName: "kube-api-access-b96j7") pod "c9446f9a-a59b-485b-be69-82419c46d6cd" (UID: "c9446f9a-a59b-485b-be69-82419c46d6cd"). InnerVolumeSpecName "kube-api-access-b96j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.600528 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9446f9a-a59b-485b-be69-82419c46d6cd" (UID: "c9446f9a-a59b-485b-be69-82419c46d6cd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.607653 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-inventory" (OuterVolumeSpecName: "inventory") pod "c9446f9a-a59b-485b-be69-82419c46d6cd" (UID: "c9446f9a-a59b-485b-be69-82419c46d6cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.656444 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.656476 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b96j7\" (UniqueName: \"kubernetes.io/projected/c9446f9a-a59b-485b-be69-82419c46d6cd-kube-api-access-b96j7\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.656488 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9446f9a-a59b-485b-be69-82419c46d6cd-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.950734 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" event={"ID":"c9446f9a-a59b-485b-be69-82419c46d6cd","Type":"ContainerDied","Data":"0ec832cbda89787949524b5e688d9788d09f570486d9684b19b0c2bd309a2c92"} Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.950801 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv" Oct 04 11:22:24 crc kubenswrapper[4758]: I1004 11:22:24.951228 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ec832cbda89787949524b5e688d9788d09f570486d9684b19b0c2bd309a2c92" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.064956 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-p8jvs"] Oct 04 11:22:25 crc kubenswrapper[4758]: E1004 11:22:25.065630 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9446f9a-a59b-485b-be69-82419c46d6cd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.065700 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9446f9a-a59b-485b-be69-82419c46d6cd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.066002 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9446f9a-a59b-485b-be69-82419c46d6cd" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.066840 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.077617 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.078215 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.078450 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.086541 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-p8jvs"] Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.087412 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.167038 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.167351 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dwxv\" (UniqueName: \"kubernetes.io/projected/03f5abcc-d8de-4c25-9664-149f1700a7fb-kube-api-access-6dwxv\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.167539 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.268666 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.268746 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dwxv\" (UniqueName: \"kubernetes.io/projected/03f5abcc-d8de-4c25-9664-149f1700a7fb-kube-api-access-6dwxv\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.268774 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.274130 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.274188 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.296864 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dwxv\" (UniqueName: \"kubernetes.io/projected/03f5abcc-d8de-4c25-9664-149f1700a7fb-kube-api-access-6dwxv\") pod \"ssh-known-hosts-edpm-deployment-p8jvs\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:25 crc kubenswrapper[4758]: I1004 11:22:25.430163 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:26 crc kubenswrapper[4758]: I1004 11:22:26.164730 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-p8jvs"] Oct 04 11:22:26 crc kubenswrapper[4758]: I1004 11:22:26.970754 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" event={"ID":"03f5abcc-d8de-4c25-9664-149f1700a7fb","Type":"ContainerStarted","Data":"f504f77b1afe7060d475ff3d489669c16f09429f859df954794616f7a3b8fc8f"} Oct 04 11:22:27 crc kubenswrapper[4758]: I1004 11:22:27.980827 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" event={"ID":"03f5abcc-d8de-4c25-9664-149f1700a7fb","Type":"ContainerStarted","Data":"ae8d4cf8f81e3c3267c881ed5a37e6c86b2aca5d76315b99e94910ca37b193c3"} Oct 04 11:22:27 crc kubenswrapper[4758]: I1004 11:22:27.997876 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" podStartSLOduration=2.302713733 podStartE2EDuration="2.997859357s" podCreationTimestamp="2025-10-04 11:22:25 +0000 UTC" firstStartedPulling="2025-10-04 11:22:26.154350535 +0000 UTC m=+1923.447001444" lastFinishedPulling="2025-10-04 11:22:26.849496149 +0000 UTC m=+1924.142147068" observedRunningTime="2025-10-04 11:22:27.995140265 +0000 UTC m=+1925.287791154" watchObservedRunningTime="2025-10-04 11:22:27.997859357 +0000 UTC m=+1925.290510246" Oct 04 11:22:35 crc kubenswrapper[4758]: I1004 11:22:35.065328 4758 generic.go:334] "Generic (PLEG): container finished" podID="03f5abcc-d8de-4c25-9664-149f1700a7fb" containerID="ae8d4cf8f81e3c3267c881ed5a37e6c86b2aca5d76315b99e94910ca37b193c3" exitCode=0 Oct 04 11:22:35 crc kubenswrapper[4758]: I1004 11:22:35.065412 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" event={"ID":"03f5abcc-d8de-4c25-9664-149f1700a7fb","Type":"ContainerDied","Data":"ae8d4cf8f81e3c3267c881ed5a37e6c86b2aca5d76315b99e94910ca37b193c3"} Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.476535 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.577777 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dwxv\" (UniqueName: \"kubernetes.io/projected/03f5abcc-d8de-4c25-9664-149f1700a7fb-kube-api-access-6dwxv\") pod \"03f5abcc-d8de-4c25-9664-149f1700a7fb\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.577834 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-ssh-key-openstack-edpm-ipam\") pod \"03f5abcc-d8de-4c25-9664-149f1700a7fb\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.577874 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-inventory-0\") pod \"03f5abcc-d8de-4c25-9664-149f1700a7fb\" (UID: \"03f5abcc-d8de-4c25-9664-149f1700a7fb\") " Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.583746 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03f5abcc-d8de-4c25-9664-149f1700a7fb-kube-api-access-6dwxv" (OuterVolumeSpecName: "kube-api-access-6dwxv") pod "03f5abcc-d8de-4c25-9664-149f1700a7fb" (UID: "03f5abcc-d8de-4c25-9664-149f1700a7fb"). InnerVolumeSpecName "kube-api-access-6dwxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.607848 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "03f5abcc-d8de-4c25-9664-149f1700a7fb" (UID: "03f5abcc-d8de-4c25-9664-149f1700a7fb"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.609311 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "03f5abcc-d8de-4c25-9664-149f1700a7fb" (UID: "03f5abcc-d8de-4c25-9664-149f1700a7fb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.680363 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dwxv\" (UniqueName: \"kubernetes.io/projected/03f5abcc-d8de-4c25-9664-149f1700a7fb-kube-api-access-6dwxv\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.680403 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:36 crc kubenswrapper[4758]: I1004 11:22:36.680415 4758 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/03f5abcc-d8de-4c25-9664-149f1700a7fb-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.089620 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" event={"ID":"03f5abcc-d8de-4c25-9664-149f1700a7fb","Type":"ContainerDied","Data":"f504f77b1afe7060d475ff3d489669c16f09429f859df954794616f7a3b8fc8f"} Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.089705 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-p8jvs" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.089721 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f504f77b1afe7060d475ff3d489669c16f09429f859df954794616f7a3b8fc8f" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.157539 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9"] Oct 04 11:22:37 crc kubenswrapper[4758]: E1004 11:22:37.157921 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03f5abcc-d8de-4c25-9664-149f1700a7fb" containerName="ssh-known-hosts-edpm-deployment" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.157935 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="03f5abcc-d8de-4c25-9664-149f1700a7fb" containerName="ssh-known-hosts-edpm-deployment" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.158144 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="03f5abcc-d8de-4c25-9664-149f1700a7fb" containerName="ssh-known-hosts-edpm-deployment" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.168722 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9"] Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.168815 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.174507 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.174553 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.174507 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.175460 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.293469 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.293650 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.293684 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczwj\" (UniqueName: \"kubernetes.io/projected/bc1d262a-55f1-46de-8584-fe90c7c82ba5-kube-api-access-pczwj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.394939 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.394987 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczwj\" (UniqueName: \"kubernetes.io/projected/bc1d262a-55f1-46de-8584-fe90c7c82ba5-kube-api-access-pczwj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.395088 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.400287 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.403685 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.417930 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczwj\" (UniqueName: \"kubernetes.io/projected/bc1d262a-55f1-46de-8584-fe90c7c82ba5-kube-api-access-pczwj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6szf9\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:37 crc kubenswrapper[4758]: I1004 11:22:37.498499 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:38 crc kubenswrapper[4758]: I1004 11:22:38.048337 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9"] Oct 04 11:22:38 crc kubenswrapper[4758]: I1004 11:22:38.097226 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" event={"ID":"bc1d262a-55f1-46de-8584-fe90c7c82ba5","Type":"ContainerStarted","Data":"63fe7c50c44dafbfc82a2fa9b430b5568de36d79446ab5c23a5b1c218f282bdd"} Oct 04 11:22:39 crc kubenswrapper[4758]: I1004 11:22:39.107792 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" event={"ID":"bc1d262a-55f1-46de-8584-fe90c7c82ba5","Type":"ContainerStarted","Data":"d51e76ca7ee9c6306f28600a83ec630408ae02587d164815968d06c1f54c2fe7"} Oct 04 11:22:47 crc kubenswrapper[4758]: I1004 11:22:47.168205 4758 generic.go:334] "Generic (PLEG): container finished" podID="bc1d262a-55f1-46de-8584-fe90c7c82ba5" containerID="d51e76ca7ee9c6306f28600a83ec630408ae02587d164815968d06c1f54c2fe7" exitCode=0 Oct 04 11:22:47 crc kubenswrapper[4758]: I1004 11:22:47.168301 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" event={"ID":"bc1d262a-55f1-46de-8584-fe90c7c82ba5","Type":"ContainerDied","Data":"d51e76ca7ee9c6306f28600a83ec630408ae02587d164815968d06c1f54c2fe7"} Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.539628 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.704710 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pczwj\" (UniqueName: \"kubernetes.io/projected/bc1d262a-55f1-46de-8584-fe90c7c82ba5-kube-api-access-pczwj\") pod \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.704872 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-inventory\") pod \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.705210 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-ssh-key\") pod \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\" (UID: \"bc1d262a-55f1-46de-8584-fe90c7c82ba5\") " Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.710531 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc1d262a-55f1-46de-8584-fe90c7c82ba5-kube-api-access-pczwj" (OuterVolumeSpecName: "kube-api-access-pczwj") pod "bc1d262a-55f1-46de-8584-fe90c7c82ba5" (UID: "bc1d262a-55f1-46de-8584-fe90c7c82ba5"). InnerVolumeSpecName "kube-api-access-pczwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.731074 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc1d262a-55f1-46de-8584-fe90c7c82ba5" (UID: "bc1d262a-55f1-46de-8584-fe90c7c82ba5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.733052 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-inventory" (OuterVolumeSpecName: "inventory") pod "bc1d262a-55f1-46de-8584-fe90c7c82ba5" (UID: "bc1d262a-55f1-46de-8584-fe90c7c82ba5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.807326 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pczwj\" (UniqueName: \"kubernetes.io/projected/bc1d262a-55f1-46de-8584-fe90c7c82ba5-kube-api-access-pczwj\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.807361 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:48 crc kubenswrapper[4758]: I1004 11:22:48.807372 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc1d262a-55f1-46de-8584-fe90c7c82ba5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.185588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" event={"ID":"bc1d262a-55f1-46de-8584-fe90c7c82ba5","Type":"ContainerDied","Data":"63fe7c50c44dafbfc82a2fa9b430b5568de36d79446ab5c23a5b1c218f282bdd"} Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.185627 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63fe7c50c44dafbfc82a2fa9b430b5568de36d79446ab5c23a5b1c218f282bdd" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.185631 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6szf9" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.338870 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds"] Oct 04 11:22:49 crc kubenswrapper[4758]: E1004 11:22:49.339347 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc1d262a-55f1-46de-8584-fe90c7c82ba5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.339372 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc1d262a-55f1-46de-8584-fe90c7c82ba5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.339695 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc1d262a-55f1-46de-8584-fe90c7c82ba5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.340643 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.346337 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.346493 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.346987 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.350881 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.353857 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds"] Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.520316 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq4v6\" (UniqueName: \"kubernetes.io/projected/ba6042e4-a924-4ddc-b91f-113c05d8ef70-kube-api-access-nq4v6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.520379 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.520409 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.622613 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq4v6\" (UniqueName: \"kubernetes.io/projected/ba6042e4-a924-4ddc-b91f-113c05d8ef70-kube-api-access-nq4v6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.623398 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.623497 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.627316 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.633057 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.637341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq4v6\" (UniqueName: \"kubernetes.io/projected/ba6042e4-a924-4ddc-b91f-113c05d8ef70-kube-api-access-nq4v6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:49 crc kubenswrapper[4758]: I1004 11:22:49.657070 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:22:50 crc kubenswrapper[4758]: I1004 11:22:50.205080 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds"] Oct 04 11:22:51 crc kubenswrapper[4758]: I1004 11:22:51.228299 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" event={"ID":"ba6042e4-a924-4ddc-b91f-113c05d8ef70","Type":"ContainerStarted","Data":"1e654fdf6c5442619ecb43886545080ed46cd7c3c5a00251a6768da0325faa95"} Oct 04 11:22:51 crc kubenswrapper[4758]: I1004 11:22:51.228701 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" event={"ID":"ba6042e4-a924-4ddc-b91f-113c05d8ef70","Type":"ContainerStarted","Data":"693fd8db9f5b5477ae2eb1c04844563f09dfdc4fc963e77d0bfa6a1de141d038"} Oct 04 11:23:01 crc kubenswrapper[4758]: I1004 11:23:01.317559 4758 generic.go:334] "Generic (PLEG): container finished" podID="ba6042e4-a924-4ddc-b91f-113c05d8ef70" containerID="1e654fdf6c5442619ecb43886545080ed46cd7c3c5a00251a6768da0325faa95" exitCode=0 Oct 04 11:23:01 crc kubenswrapper[4758]: I1004 11:23:01.317659 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" event={"ID":"ba6042e4-a924-4ddc-b91f-113c05d8ef70","Type":"ContainerDied","Data":"1e654fdf6c5442619ecb43886545080ed46cd7c3c5a00251a6768da0325faa95"} Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.727241 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.869460 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-ssh-key\") pod \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.869843 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq4v6\" (UniqueName: \"kubernetes.io/projected/ba6042e4-a924-4ddc-b91f-113c05d8ef70-kube-api-access-nq4v6\") pod \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.869918 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-inventory\") pod \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\" (UID: \"ba6042e4-a924-4ddc-b91f-113c05d8ef70\") " Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.890333 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6042e4-a924-4ddc-b91f-113c05d8ef70-kube-api-access-nq4v6" (OuterVolumeSpecName: "kube-api-access-nq4v6") pod "ba6042e4-a924-4ddc-b91f-113c05d8ef70" (UID: "ba6042e4-a924-4ddc-b91f-113c05d8ef70"). InnerVolumeSpecName "kube-api-access-nq4v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.916278 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-inventory" (OuterVolumeSpecName: "inventory") pod "ba6042e4-a924-4ddc-b91f-113c05d8ef70" (UID: "ba6042e4-a924-4ddc-b91f-113c05d8ef70"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.916325 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba6042e4-a924-4ddc-b91f-113c05d8ef70" (UID: "ba6042e4-a924-4ddc-b91f-113c05d8ef70"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.972115 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.972160 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq4v6\" (UniqueName: \"kubernetes.io/projected/ba6042e4-a924-4ddc-b91f-113c05d8ef70-kube-api-access-nq4v6\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:02 crc kubenswrapper[4758]: I1004 11:23:02.972175 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6042e4-a924-4ddc-b91f-113c05d8ef70-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.352793 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" event={"ID":"ba6042e4-a924-4ddc-b91f-113c05d8ef70","Type":"ContainerDied","Data":"693fd8db9f5b5477ae2eb1c04844563f09dfdc4fc963e77d0bfa6a1de141d038"} Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.352839 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="693fd8db9f5b5477ae2eb1c04844563f09dfdc4fc963e77d0bfa6a1de141d038" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.352957 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.447382 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp"] Oct 04 11:23:03 crc kubenswrapper[4758]: E1004 11:23:03.447777 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6042e4-a924-4ddc-b91f-113c05d8ef70" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.447794 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6042e4-a924-4ddc-b91f-113c05d8ef70" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.448028 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6042e4-a924-4ddc-b91f-113c05d8ef70" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.448775 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.451005 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.451061 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.451188 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.451372 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.452785 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.453206 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.453334 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.453558 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.469474 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp"] Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594085 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594161 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594194 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594235 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594256 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594307 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqfgm\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-kube-api-access-qqfgm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594376 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594402 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594451 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594478 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594495 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.594545 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696066 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696167 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696197 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696234 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696276 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696301 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696325 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696365 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696390 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696449 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqfgm\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-kube-api-access-qqfgm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696471 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696508 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696531 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.696561 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.706322 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.706502 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.713061 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.718716 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.722114 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.722945 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.723746 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.724022 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.728762 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.729589 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.730843 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.734629 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.735831 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqfgm\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-kube-api-access-qqfgm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.752677 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:03 crc kubenswrapper[4758]: I1004 11:23:03.769493 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:04 crc kubenswrapper[4758]: I1004 11:23:04.425290 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp"] Oct 04 11:23:04 crc kubenswrapper[4758]: W1004 11:23:04.435326 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf631c48_0a7f_41cf_911c_c34c42b30558.slice/crio-1bb56e5c4acd7148b5affb9a20e4008fd9e0a13a22671522b2930e96608dde7f WatchSource:0}: Error finding container 1bb56e5c4acd7148b5affb9a20e4008fd9e0a13a22671522b2930e96608dde7f: Status 404 returned error can't find the container with id 1bb56e5c4acd7148b5affb9a20e4008fd9e0a13a22671522b2930e96608dde7f Oct 04 11:23:05 crc kubenswrapper[4758]: I1004 11:23:05.372640 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" event={"ID":"bf631c48-0a7f-41cf-911c-c34c42b30558","Type":"ContainerStarted","Data":"7614bfb90cf309c7733e329934a84c4ca87771b5a1eb80ed72b976690fbfa8b1"} Oct 04 11:23:05 crc kubenswrapper[4758]: I1004 11:23:05.373031 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" event={"ID":"bf631c48-0a7f-41cf-911c-c34c42b30558","Type":"ContainerStarted","Data":"1bb56e5c4acd7148b5affb9a20e4008fd9e0a13a22671522b2930e96608dde7f"} Oct 04 11:23:05 crc kubenswrapper[4758]: I1004 11:23:05.404419 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" podStartSLOduration=1.934770246 podStartE2EDuration="2.404399038s" podCreationTimestamp="2025-10-04 11:23:03 +0000 UTC" firstStartedPulling="2025-10-04 11:23:04.438694531 +0000 UTC m=+1961.731345420" lastFinishedPulling="2025-10-04 11:23:04.908323323 +0000 UTC m=+1962.200974212" observedRunningTime="2025-10-04 11:23:05.397562006 +0000 UTC m=+1962.690212915" watchObservedRunningTime="2025-10-04 11:23:05.404399038 +0000 UTC m=+1962.697049927" Oct 04 11:23:48 crc kubenswrapper[4758]: I1004 11:23:48.759537 4758 generic.go:334] "Generic (PLEG): container finished" podID="bf631c48-0a7f-41cf-911c-c34c42b30558" containerID="7614bfb90cf309c7733e329934a84c4ca87771b5a1eb80ed72b976690fbfa8b1" exitCode=0 Oct 04 11:23:48 crc kubenswrapper[4758]: I1004 11:23:48.759566 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" event={"ID":"bf631c48-0a7f-41cf-911c-c34c42b30558","Type":"ContainerDied","Data":"7614bfb90cf309c7733e329934a84c4ca87771b5a1eb80ed72b976690fbfa8b1"} Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.185670 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.309967 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqfgm\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-kube-api-access-qqfgm\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310148 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-repo-setup-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310173 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ovn-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310206 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-bootstrap-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310237 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310259 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-telemetry-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310338 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-nova-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310360 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310404 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-libvirt-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310420 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-inventory\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310438 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310483 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-ovn-default-certs-0\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310507 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-neutron-metadata-combined-ca-bundle\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.310534 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ssh-key\") pod \"bf631c48-0a7f-41cf-911c-c34c42b30558\" (UID: \"bf631c48-0a7f-41cf-911c-c34c42b30558\") " Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.318093 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-kube-api-access-qqfgm" (OuterVolumeSpecName: "kube-api-access-qqfgm") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "kube-api-access-qqfgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.318904 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.321279 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.321345 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.322575 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.323918 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.323943 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.325340 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.327693 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.328554 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.331403 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.332234 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.345447 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.347721 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-inventory" (OuterVolumeSpecName: "inventory") pod "bf631c48-0a7f-41cf-911c-c34c42b30558" (UID: "bf631c48-0a7f-41cf-911c-c34c42b30558"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413219 4758 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413260 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413273 4758 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413287 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413307 4758 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413324 4758 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413341 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413360 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413380 4758 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413399 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413415 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413432 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413450 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf631c48-0a7f-41cf-911c-c34c42b30558-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.413467 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqfgm\" (UniqueName: \"kubernetes.io/projected/bf631c48-0a7f-41cf-911c-c34c42b30558-kube-api-access-qqfgm\") on node \"crc\" DevicePath \"\"" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.779955 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" event={"ID":"bf631c48-0a7f-41cf-911c-c34c42b30558","Type":"ContainerDied","Data":"1bb56e5c4acd7148b5affb9a20e4008fd9e0a13a22671522b2930e96608dde7f"} Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.779991 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bb56e5c4acd7148b5affb9a20e4008fd9e0a13a22671522b2930e96608dde7f" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.780007 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.958853 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5"] Oct 04 11:23:50 crc kubenswrapper[4758]: E1004 11:23:50.959302 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf631c48-0a7f-41cf-911c-c34c42b30558" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.959325 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf631c48-0a7f-41cf-911c-c34c42b30558" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.959567 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf631c48-0a7f-41cf-911c-c34c42b30558" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.960387 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.962497 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.963624 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.963790 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.964371 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.973019 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:23:50 crc kubenswrapper[4758]: I1004 11:23:50.983356 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5"] Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.026371 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24kpx\" (UniqueName: \"kubernetes.io/projected/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-kube-api-access-24kpx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.026441 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.026472 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.026727 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.026882 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.128711 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.128829 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.128910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24kpx\" (UniqueName: \"kubernetes.io/projected/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-kube-api-access-24kpx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.128938 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.128967 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.129905 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.134029 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.134036 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.135864 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.156561 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24kpx\" (UniqueName: \"kubernetes.io/projected/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-kube-api-access-24kpx\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qdvb5\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.288339 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:23:51 crc kubenswrapper[4758]: I1004 11:23:51.853127 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5"] Oct 04 11:23:52 crc kubenswrapper[4758]: I1004 11:23:52.798800 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" event={"ID":"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3","Type":"ContainerStarted","Data":"db78bb66278ec9595971db2aff102e05b22d2a64c5a6a82287103aa1b4712721"} Oct 04 11:23:52 crc kubenswrapper[4758]: I1004 11:23:52.799153 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" event={"ID":"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3","Type":"ContainerStarted","Data":"7a1bfc83118f1764c9ec3f1db845e513b35f5ef1e0a670b64e5c0dce79b52f5e"} Oct 04 11:23:52 crc kubenswrapper[4758]: I1004 11:23:52.821193 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" podStartSLOduration=2.268728367 podStartE2EDuration="2.821175263s" podCreationTimestamp="2025-10-04 11:23:50 +0000 UTC" firstStartedPulling="2025-10-04 11:23:51.865039681 +0000 UTC m=+2009.157690570" lastFinishedPulling="2025-10-04 11:23:52.417486537 +0000 UTC m=+2009.710137466" observedRunningTime="2025-10-04 11:23:52.81996205 +0000 UTC m=+2010.112612959" watchObservedRunningTime="2025-10-04 11:23:52.821175263 +0000 UTC m=+2010.113826162" Oct 04 11:24:31 crc kubenswrapper[4758]: I1004 11:24:31.249848 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:24:31 crc kubenswrapper[4758]: I1004 11:24:31.250366 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:25:01 crc kubenswrapper[4758]: I1004 11:25:01.249729 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:25:01 crc kubenswrapper[4758]: I1004 11:25:01.250421 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:25:08 crc kubenswrapper[4758]: I1004 11:25:08.485690 4758 generic.go:334] "Generic (PLEG): container finished" podID="dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" containerID="db78bb66278ec9595971db2aff102e05b22d2a64c5a6a82287103aa1b4712721" exitCode=0 Oct 04 11:25:08 crc kubenswrapper[4758]: I1004 11:25:08.485790 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" event={"ID":"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3","Type":"ContainerDied","Data":"db78bb66278ec9595971db2aff102e05b22d2a64c5a6a82287103aa1b4712721"} Oct 04 11:25:09 crc kubenswrapper[4758]: I1004 11:25:09.911828 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.063567 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovn-combined-ca-bundle\") pod \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.063694 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ssh-key\") pod \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.064431 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24kpx\" (UniqueName: \"kubernetes.io/projected/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-kube-api-access-24kpx\") pod \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.064471 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-inventory\") pod \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.064499 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovncontroller-config-0\") pod \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\" (UID: \"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3\") " Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.068762 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" (UID: "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.069503 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-kube-api-access-24kpx" (OuterVolumeSpecName: "kube-api-access-24kpx") pod "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" (UID: "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3"). InnerVolumeSpecName "kube-api-access-24kpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.088267 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" (UID: "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.095573 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" (UID: "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.115733 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-inventory" (OuterVolumeSpecName: "inventory") pod "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" (UID: "dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.166763 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.166793 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24kpx\" (UniqueName: \"kubernetes.io/projected/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-kube-api-access-24kpx\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.166804 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.166814 4758 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.166823 4758 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.507584 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" event={"ID":"dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3","Type":"ContainerDied","Data":"7a1bfc83118f1764c9ec3f1db845e513b35f5ef1e0a670b64e5c0dce79b52f5e"} Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.507618 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a1bfc83118f1764c9ec3f1db845e513b35f5ef1e0a670b64e5c0dce79b52f5e" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.507679 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qdvb5" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.622921 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj"] Oct 04 11:25:10 crc kubenswrapper[4758]: E1004 11:25:10.623290 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.623305 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.623486 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.624073 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.627979 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.628180 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.628284 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.635360 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.636335 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.636586 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.644344 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj"] Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.778378 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.778453 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k6ht\" (UniqueName: \"kubernetes.io/projected/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-kube-api-access-5k6ht\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.778503 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.778728 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.778921 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.778991 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.880930 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.880999 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k6ht\" (UniqueName: \"kubernetes.io/projected/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-kube-api-access-5k6ht\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.881047 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.881080 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.881140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.881168 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.885286 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.886187 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.886720 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.887530 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.888635 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.922496 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k6ht\" (UniqueName: \"kubernetes.io/projected/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-kube-api-access-5k6ht\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:10 crc kubenswrapper[4758]: I1004 11:25:10.941978 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:25:11 crc kubenswrapper[4758]: I1004 11:25:11.448372 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj"] Oct 04 11:25:11 crc kubenswrapper[4758]: W1004 11:25:11.457329 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bdca8ac_44ee_4fc1_8b08_7f592ca42a4e.slice/crio-ee1345af1d9ef53eed1418a58bee9a96e421967b915dbc0e59421e355dc818dc WatchSource:0}: Error finding container ee1345af1d9ef53eed1418a58bee9a96e421967b915dbc0e59421e355dc818dc: Status 404 returned error can't find the container with id ee1345af1d9ef53eed1418a58bee9a96e421967b915dbc0e59421e355dc818dc Oct 04 11:25:11 crc kubenswrapper[4758]: I1004 11:25:11.516240 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" event={"ID":"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e","Type":"ContainerStarted","Data":"ee1345af1d9ef53eed1418a58bee9a96e421967b915dbc0e59421e355dc818dc"} Oct 04 11:25:12 crc kubenswrapper[4758]: I1004 11:25:12.526271 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" event={"ID":"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e","Type":"ContainerStarted","Data":"a4f8e16ad02a861ca12433a36d31c627cefa4a194714121161da31798e858f37"} Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.249154 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.249812 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.249868 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.250643 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8fb28a3d28b026315f63e9ed9f93326a8b51ed6380ad4d69c7930ab2c772926d"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.250705 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://8fb28a3d28b026315f63e9ed9f93326a8b51ed6380ad4d69c7930ab2c772926d" gracePeriod=600 Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.710010 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="8fb28a3d28b026315f63e9ed9f93326a8b51ed6380ad4d69c7930ab2c772926d" exitCode=0 Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.710082 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"8fb28a3d28b026315f63e9ed9f93326a8b51ed6380ad4d69c7930ab2c772926d"} Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.710473 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888"} Oct 04 11:25:31 crc kubenswrapper[4758]: I1004 11:25:31.710502 4758 scope.go:117] "RemoveContainer" containerID="330695cf9e1b84eb428328adcc8cc285d758549665f67c0920fc9d768b313f55" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.122204 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" podStartSLOduration=21.572975452 podStartE2EDuration="22.122171531s" podCreationTimestamp="2025-10-04 11:25:10 +0000 UTC" firstStartedPulling="2025-10-04 11:25:11.459240819 +0000 UTC m=+2088.751891708" lastFinishedPulling="2025-10-04 11:25:12.008436898 +0000 UTC m=+2089.301087787" observedRunningTime="2025-10-04 11:25:12.558845169 +0000 UTC m=+2089.851496088" watchObservedRunningTime="2025-10-04 11:25:32.122171531 +0000 UTC m=+2109.414822430" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.122996 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8vrwh"] Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.125169 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.140198 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vrwh"] Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.198302 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-utilities\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.198454 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-catalog-content\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.198512 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffjdw\" (UniqueName: \"kubernetes.io/projected/e94f57f3-78e6-4c95-8c1a-84088c7285dc-kube-api-access-ffjdw\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.300805 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-catalog-content\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.300936 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffjdw\" (UniqueName: \"kubernetes.io/projected/e94f57f3-78e6-4c95-8c1a-84088c7285dc-kube-api-access-ffjdw\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.301013 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-utilities\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.301465 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-catalog-content\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.301569 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-utilities\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.323056 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffjdw\" (UniqueName: \"kubernetes.io/projected/e94f57f3-78e6-4c95-8c1a-84088c7285dc-kube-api-access-ffjdw\") pod \"certified-operators-8vrwh\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.450465 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:32 crc kubenswrapper[4758]: I1004 11:25:32.987650 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vrwh"] Oct 04 11:25:33 crc kubenswrapper[4758]: I1004 11:25:33.760970 4758 generic.go:334] "Generic (PLEG): container finished" podID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerID="62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb" exitCode=0 Oct 04 11:25:33 crc kubenswrapper[4758]: I1004 11:25:33.761020 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerDied","Data":"62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb"} Oct 04 11:25:33 crc kubenswrapper[4758]: I1004 11:25:33.761639 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerStarted","Data":"e3d49ea4e002968fb60e0356cf006633a87fd5c1383e080b5ca68631ca014b0e"} Oct 04 11:25:34 crc kubenswrapper[4758]: I1004 11:25:34.771610 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerStarted","Data":"0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59"} Oct 04 11:25:35 crc kubenswrapper[4758]: I1004 11:25:35.784216 4758 generic.go:334] "Generic (PLEG): container finished" podID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerID="0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59" exitCode=0 Oct 04 11:25:35 crc kubenswrapper[4758]: I1004 11:25:35.784278 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerDied","Data":"0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59"} Oct 04 11:25:36 crc kubenswrapper[4758]: I1004 11:25:36.798025 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerStarted","Data":"6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43"} Oct 04 11:25:36 crc kubenswrapper[4758]: I1004 11:25:36.823797 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8vrwh" podStartSLOduration=2.271088932 podStartE2EDuration="4.823765504s" podCreationTimestamp="2025-10-04 11:25:32 +0000 UTC" firstStartedPulling="2025-10-04 11:25:33.765913315 +0000 UTC m=+2111.058564214" lastFinishedPulling="2025-10-04 11:25:36.318589897 +0000 UTC m=+2113.611240786" observedRunningTime="2025-10-04 11:25:36.820881707 +0000 UTC m=+2114.113532626" watchObservedRunningTime="2025-10-04 11:25:36.823765504 +0000 UTC m=+2114.116416383" Oct 04 11:25:42 crc kubenswrapper[4758]: I1004 11:25:42.451409 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:42 crc kubenswrapper[4758]: I1004 11:25:42.451992 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:42 crc kubenswrapper[4758]: I1004 11:25:42.537205 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:42 crc kubenswrapper[4758]: I1004 11:25:42.890873 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:42 crc kubenswrapper[4758]: I1004 11:25:42.950052 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vrwh"] Oct 04 11:25:44 crc kubenswrapper[4758]: I1004 11:25:44.870599 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8vrwh" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="registry-server" containerID="cri-o://6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43" gracePeriod=2 Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.291631 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.334218 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-catalog-content\") pod \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.334323 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffjdw\" (UniqueName: \"kubernetes.io/projected/e94f57f3-78e6-4c95-8c1a-84088c7285dc-kube-api-access-ffjdw\") pod \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.334446 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-utilities\") pod \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\" (UID: \"e94f57f3-78e6-4c95-8c1a-84088c7285dc\") " Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.335443 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-utilities" (OuterVolumeSpecName: "utilities") pod "e94f57f3-78e6-4c95-8c1a-84088c7285dc" (UID: "e94f57f3-78e6-4c95-8c1a-84088c7285dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.340665 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94f57f3-78e6-4c95-8c1a-84088c7285dc-kube-api-access-ffjdw" (OuterVolumeSpecName: "kube-api-access-ffjdw") pod "e94f57f3-78e6-4c95-8c1a-84088c7285dc" (UID: "e94f57f3-78e6-4c95-8c1a-84088c7285dc"). InnerVolumeSpecName "kube-api-access-ffjdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.391917 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e94f57f3-78e6-4c95-8c1a-84088c7285dc" (UID: "e94f57f3-78e6-4c95-8c1a-84088c7285dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.436226 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.436251 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94f57f3-78e6-4c95-8c1a-84088c7285dc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.436263 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffjdw\" (UniqueName: \"kubernetes.io/projected/e94f57f3-78e6-4c95-8c1a-84088c7285dc-kube-api-access-ffjdw\") on node \"crc\" DevicePath \"\"" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.879743 4758 generic.go:334] "Generic (PLEG): container finished" podID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerID="6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43" exitCode=0 Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.879784 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerDied","Data":"6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43"} Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.879793 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vrwh" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.879809 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vrwh" event={"ID":"e94f57f3-78e6-4c95-8c1a-84088c7285dc","Type":"ContainerDied","Data":"e3d49ea4e002968fb60e0356cf006633a87fd5c1383e080b5ca68631ca014b0e"} Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.879829 4758 scope.go:117] "RemoveContainer" containerID="6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.896974 4758 scope.go:117] "RemoveContainer" containerID="0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.921192 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vrwh"] Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.924665 4758 scope.go:117] "RemoveContainer" containerID="62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.927587 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8vrwh"] Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.964909 4758 scope.go:117] "RemoveContainer" containerID="6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43" Oct 04 11:25:45 crc kubenswrapper[4758]: E1004 11:25:45.965403 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43\": container with ID starting with 6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43 not found: ID does not exist" containerID="6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.965452 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43"} err="failed to get container status \"6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43\": rpc error: code = NotFound desc = could not find container \"6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43\": container with ID starting with 6d57c3b6a454ed85ad5f7bc464916b7817e4ec9277ef34b643b662362b36fc43 not found: ID does not exist" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.965480 4758 scope.go:117] "RemoveContainer" containerID="0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59" Oct 04 11:25:45 crc kubenswrapper[4758]: E1004 11:25:45.965791 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59\": container with ID starting with 0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59 not found: ID does not exist" containerID="0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.965823 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59"} err="failed to get container status \"0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59\": rpc error: code = NotFound desc = could not find container \"0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59\": container with ID starting with 0ae9738801070f007e6b1a01dfba49a290ae11d68b2b5a3fab323a4f39c68a59 not found: ID does not exist" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.965845 4758 scope.go:117] "RemoveContainer" containerID="62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb" Oct 04 11:25:45 crc kubenswrapper[4758]: E1004 11:25:45.966094 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb\": container with ID starting with 62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb not found: ID does not exist" containerID="62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb" Oct 04 11:25:45 crc kubenswrapper[4758]: I1004 11:25:45.966150 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb"} err="failed to get container status \"62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb\": rpc error: code = NotFound desc = could not find container \"62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb\": container with ID starting with 62446cd9f955682855bda0c393d6eb997421d58154155b4e219c69ed7e64fbfb not found: ID does not exist" Oct 04 11:25:47 crc kubenswrapper[4758]: I1004 11:25:47.334577 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" path="/var/lib/kubelet/pods/e94f57f3-78e6-4c95-8c1a-84088c7285dc/volumes" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.923215 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lt4lm"] Oct 04 11:25:48 crc kubenswrapper[4758]: E1004 11:25:48.923727 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="extract-content" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.923743 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="extract-content" Oct 04 11:25:48 crc kubenswrapper[4758]: E1004 11:25:48.923762 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="extract-utilities" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.923771 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="extract-utilities" Oct 04 11:25:48 crc kubenswrapper[4758]: E1004 11:25:48.923800 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="registry-server" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.923809 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="registry-server" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.924047 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94f57f3-78e6-4c95-8c1a-84088c7285dc" containerName="registry-server" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.925823 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:48 crc kubenswrapper[4758]: I1004 11:25:48.943073 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lt4lm"] Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.009045 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-catalog-content\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.009162 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-utilities\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.009214 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkbsd\" (UniqueName: \"kubernetes.io/projected/a3ba291f-bd05-443b-93e9-5339d29d839d-kube-api-access-mkbsd\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.110891 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-utilities\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.110948 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkbsd\" (UniqueName: \"kubernetes.io/projected/a3ba291f-bd05-443b-93e9-5339d29d839d-kube-api-access-mkbsd\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.111052 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-catalog-content\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.111497 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-catalog-content\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.111700 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-utilities\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.130810 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkbsd\" (UniqueName: \"kubernetes.io/projected/a3ba291f-bd05-443b-93e9-5339d29d839d-kube-api-access-mkbsd\") pod \"community-operators-lt4lm\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.249412 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.663457 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lt4lm"] Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.930929 4758 generic.go:334] "Generic (PLEG): container finished" podID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerID="de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232" exitCode=0 Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.930979 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerDied","Data":"de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232"} Oct 04 11:25:49 crc kubenswrapper[4758]: I1004 11:25:49.932454 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerStarted","Data":"199fa12de605ae25aad043f97a602b712f2c5b17bfbaaf1a3a54446e03daa833"} Oct 04 11:25:50 crc kubenswrapper[4758]: I1004 11:25:50.941742 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerStarted","Data":"ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880"} Oct 04 11:25:51 crc kubenswrapper[4758]: I1004 11:25:51.950774 4758 generic.go:334] "Generic (PLEG): container finished" podID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerID="ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880" exitCode=0 Oct 04 11:25:51 crc kubenswrapper[4758]: I1004 11:25:51.950983 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerDied","Data":"ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880"} Oct 04 11:25:52 crc kubenswrapper[4758]: I1004 11:25:52.962615 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerStarted","Data":"da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a"} Oct 04 11:25:53 crc kubenswrapper[4758]: I1004 11:25:53.010410 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lt4lm" podStartSLOduration=2.356093905 podStartE2EDuration="5.01039204s" podCreationTimestamp="2025-10-04 11:25:48 +0000 UTC" firstStartedPulling="2025-10-04 11:25:49.934280427 +0000 UTC m=+2127.226931306" lastFinishedPulling="2025-10-04 11:25:52.588578552 +0000 UTC m=+2129.881229441" observedRunningTime="2025-10-04 11:25:53.009191508 +0000 UTC m=+2130.301842397" watchObservedRunningTime="2025-10-04 11:25:53.01039204 +0000 UTC m=+2130.303042929" Oct 04 11:25:59 crc kubenswrapper[4758]: I1004 11:25:59.249989 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:59 crc kubenswrapper[4758]: I1004 11:25:59.251562 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:25:59 crc kubenswrapper[4758]: I1004 11:25:59.339201 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:26:00 crc kubenswrapper[4758]: I1004 11:26:00.073650 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:26:00 crc kubenswrapper[4758]: I1004 11:26:00.116692 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lt4lm"] Oct 04 11:26:01 crc kubenswrapper[4758]: I1004 11:26:01.994053 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m8r6s"] Oct 04 11:26:01 crc kubenswrapper[4758]: I1004 11:26:01.997079 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.014669 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m8r6s"] Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.049327 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lt4lm" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="registry-server" containerID="cri-o://da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a" gracePeriod=2 Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.055498 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-catalog-content\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.055577 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-utilities\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.055605 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qkf\" (UniqueName: \"kubernetes.io/projected/7d805fd7-5d93-4b4f-831e-446e51a318a9-kube-api-access-w5qkf\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.157265 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-catalog-content\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.157326 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-utilities\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.157344 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qkf\" (UniqueName: \"kubernetes.io/projected/7d805fd7-5d93-4b4f-831e-446e51a318a9-kube-api-access-w5qkf\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.158071 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-catalog-content\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.158290 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-utilities\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.178950 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qkf\" (UniqueName: \"kubernetes.io/projected/7d805fd7-5d93-4b4f-831e-446e51a318a9-kube-api-access-w5qkf\") pod \"redhat-operators-m8r6s\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.329164 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.545940 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.668944 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-utilities\") pod \"a3ba291f-bd05-443b-93e9-5339d29d839d\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.669237 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkbsd\" (UniqueName: \"kubernetes.io/projected/a3ba291f-bd05-443b-93e9-5339d29d839d-kube-api-access-mkbsd\") pod \"a3ba291f-bd05-443b-93e9-5339d29d839d\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.669295 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-catalog-content\") pod \"a3ba291f-bd05-443b-93e9-5339d29d839d\" (UID: \"a3ba291f-bd05-443b-93e9-5339d29d839d\") " Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.670235 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-utilities" (OuterVolumeSpecName: "utilities") pod "a3ba291f-bd05-443b-93e9-5339d29d839d" (UID: "a3ba291f-bd05-443b-93e9-5339d29d839d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.682770 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3ba291f-bd05-443b-93e9-5339d29d839d-kube-api-access-mkbsd" (OuterVolumeSpecName: "kube-api-access-mkbsd") pod "a3ba291f-bd05-443b-93e9-5339d29d839d" (UID: "a3ba291f-bd05-443b-93e9-5339d29d839d"). InnerVolumeSpecName "kube-api-access-mkbsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.707860 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m8r6s"] Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.748973 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3ba291f-bd05-443b-93e9-5339d29d839d" (UID: "a3ba291f-bd05-443b-93e9-5339d29d839d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.772507 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.772530 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3ba291f-bd05-443b-93e9-5339d29d839d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:02 crc kubenswrapper[4758]: I1004 11:26:02.772541 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkbsd\" (UniqueName: \"kubernetes.io/projected/a3ba291f-bd05-443b-93e9-5339d29d839d-kube-api-access-mkbsd\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.085714 4758 generic.go:334] "Generic (PLEG): container finished" podID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerID="da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a" exitCode=0 Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.085762 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lt4lm" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.085780 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerDied","Data":"da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a"} Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.086012 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lt4lm" event={"ID":"a3ba291f-bd05-443b-93e9-5339d29d839d","Type":"ContainerDied","Data":"199fa12de605ae25aad043f97a602b712f2c5b17bfbaaf1a3a54446e03daa833"} Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.086045 4758 scope.go:117] "RemoveContainer" containerID="da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.097163 4758 generic.go:334] "Generic (PLEG): container finished" podID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerID="ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292" exitCode=0 Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.097204 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerDied","Data":"ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292"} Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.097238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerStarted","Data":"98d9ac0ccc1325319d7cced84d059e8aaa4e8c3064de50e0f996b6d87f3a4d35"} Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.119623 4758 scope.go:117] "RemoveContainer" containerID="ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.139213 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lt4lm"] Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.149838 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lt4lm"] Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.171274 4758 scope.go:117] "RemoveContainer" containerID="de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.217754 4758 scope.go:117] "RemoveContainer" containerID="da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a" Oct 04 11:26:03 crc kubenswrapper[4758]: E1004 11:26:03.227680 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a\": container with ID starting with da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a not found: ID does not exist" containerID="da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.227745 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a"} err="failed to get container status \"da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a\": rpc error: code = NotFound desc = could not find container \"da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a\": container with ID starting with da8a647e5db61f1a4ef55303d8276c2e40fd11ab1b2c90869960005d60eaee3a not found: ID does not exist" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.227771 4758 scope.go:117] "RemoveContainer" containerID="ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880" Oct 04 11:26:03 crc kubenswrapper[4758]: E1004 11:26:03.228251 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880\": container with ID starting with ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880 not found: ID does not exist" containerID="ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.228295 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880"} err="failed to get container status \"ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880\": rpc error: code = NotFound desc = could not find container \"ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880\": container with ID starting with ae8b5e7083714e4918fa9f8d04ad4512fe6560c5cb9c6e57f9e5d3a779860880 not found: ID does not exist" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.228321 4758 scope.go:117] "RemoveContainer" containerID="de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232" Oct 04 11:26:03 crc kubenswrapper[4758]: E1004 11:26:03.228832 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232\": container with ID starting with de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232 not found: ID does not exist" containerID="de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.228872 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232"} err="failed to get container status \"de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232\": rpc error: code = NotFound desc = could not find container \"de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232\": container with ID starting with de7f96a5dd684a0f48ad88b782925961b597ca6e15eb96c2b8918143961ef232 not found: ID does not exist" Oct 04 11:26:03 crc kubenswrapper[4758]: I1004 11:26:03.334807 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" path="/var/lib/kubelet/pods/a3ba291f-bd05-443b-93e9-5339d29d839d/volumes" Oct 04 11:26:04 crc kubenswrapper[4758]: I1004 11:26:04.109443 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerStarted","Data":"3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d"} Oct 04 11:26:08 crc kubenswrapper[4758]: I1004 11:26:08.150275 4758 generic.go:334] "Generic (PLEG): container finished" podID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerID="3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d" exitCode=0 Oct 04 11:26:08 crc kubenswrapper[4758]: I1004 11:26:08.150434 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerDied","Data":"3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d"} Oct 04 11:26:08 crc kubenswrapper[4758]: I1004 11:26:08.154641 4758 generic.go:334] "Generic (PLEG): container finished" podID="2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" containerID="a4f8e16ad02a861ca12433a36d31c627cefa4a194714121161da31798e858f37" exitCode=0 Oct 04 11:26:08 crc kubenswrapper[4758]: I1004 11:26:08.154677 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" event={"ID":"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e","Type":"ContainerDied","Data":"a4f8e16ad02a861ca12433a36d31c627cefa4a194714121161da31798e858f37"} Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.162897 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerStarted","Data":"75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e"} Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.607071 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.712360 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-ssh-key\") pod \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.712952 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-nova-metadata-neutron-config-0\") pod \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.713209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k6ht\" (UniqueName: \"kubernetes.io/projected/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-kube-api-access-5k6ht\") pod \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.713363 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.713455 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-metadata-combined-ca-bundle\") pod \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.713578 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-inventory\") pod \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\" (UID: \"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e\") " Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.731492 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" (UID: "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.739477 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-kube-api-access-5k6ht" (OuterVolumeSpecName: "kube-api-access-5k6ht") pod "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" (UID: "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e"). InnerVolumeSpecName "kube-api-access-5k6ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.741287 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" (UID: "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.741633 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" (UID: "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.744323 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" (UID: "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.752870 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-inventory" (OuterVolumeSpecName: "inventory") pod "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" (UID: "2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.816021 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.816142 4758 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.816208 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.816272 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.816332 4758 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:09 crc kubenswrapper[4758]: I1004 11:26:09.816386 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k6ht\" (UniqueName: \"kubernetes.io/projected/2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e-kube-api-access-5k6ht\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.173765 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.173813 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj" event={"ID":"2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e","Type":"ContainerDied","Data":"ee1345af1d9ef53eed1418a58bee9a96e421967b915dbc0e59421e355dc818dc"} Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.173837 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee1345af1d9ef53eed1418a58bee9a96e421967b915dbc0e59421e355dc818dc" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.200640 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m8r6s" podStartSLOduration=3.662053349 podStartE2EDuration="9.200624904s" podCreationTimestamp="2025-10-04 11:26:01 +0000 UTC" firstStartedPulling="2025-10-04 11:26:03.101368595 +0000 UTC m=+2140.394019484" lastFinishedPulling="2025-10-04 11:26:08.63994015 +0000 UTC m=+2145.932591039" observedRunningTime="2025-10-04 11:26:10.195494697 +0000 UTC m=+2147.488145586" watchObservedRunningTime="2025-10-04 11:26:10.200624904 +0000 UTC m=+2147.493275793" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340295 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl"] Oct 04 11:26:10 crc kubenswrapper[4758]: E1004 11:26:10.340639 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340656 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 11:26:10 crc kubenswrapper[4758]: E1004 11:26:10.340673 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="extract-content" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340680 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="extract-content" Oct 04 11:26:10 crc kubenswrapper[4758]: E1004 11:26:10.340695 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="registry-server" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340701 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="registry-server" Oct 04 11:26:10 crc kubenswrapper[4758]: E1004 11:26:10.340711 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="extract-utilities" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340717 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="extract-utilities" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340887 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3ba291f-bd05-443b-93e9-5339d29d839d" containerName="registry-server" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.340914 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.341483 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.350213 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl"] Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.351569 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.351623 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.351993 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.352229 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.354183 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.426744 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.426824 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.427772 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.427827 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5d2v\" (UniqueName: \"kubernetes.io/projected/6ec33929-038a-47ec-9240-42f88c26168e-kube-api-access-p5d2v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.427891 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.529638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.529701 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.529744 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.529819 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5d2v\" (UniqueName: \"kubernetes.io/projected/6ec33929-038a-47ec-9240-42f88c26168e-kube-api-access-p5d2v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.529860 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.534552 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.534966 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.536037 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.547411 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.551610 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5d2v\" (UniqueName: \"kubernetes.io/projected/6ec33929-038a-47ec-9240-42f88c26168e-kube-api-access-p5d2v\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m97jl\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:10 crc kubenswrapper[4758]: I1004 11:26:10.702014 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:26:11 crc kubenswrapper[4758]: I1004 11:26:11.236739 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl"] Oct 04 11:26:11 crc kubenswrapper[4758]: W1004 11:26:11.246738 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ec33929_038a_47ec_9240_42f88c26168e.slice/crio-9f13545d1c8ac3bc88d595dec94526881c77fed0993e910c07fe4c6f24260904 WatchSource:0}: Error finding container 9f13545d1c8ac3bc88d595dec94526881c77fed0993e910c07fe4c6f24260904: Status 404 returned error can't find the container with id 9f13545d1c8ac3bc88d595dec94526881c77fed0993e910c07fe4c6f24260904 Oct 04 11:26:12 crc kubenswrapper[4758]: I1004 11:26:12.190872 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" event={"ID":"6ec33929-038a-47ec-9240-42f88c26168e","Type":"ContainerStarted","Data":"28c85d8ef1e3813c224ecd2eb8f60be39043a06a89185ab7e33186bc9514b416"} Oct 04 11:26:12 crc kubenswrapper[4758]: I1004 11:26:12.191726 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" event={"ID":"6ec33929-038a-47ec-9240-42f88c26168e","Type":"ContainerStarted","Data":"9f13545d1c8ac3bc88d595dec94526881c77fed0993e910c07fe4c6f24260904"} Oct 04 11:26:12 crc kubenswrapper[4758]: I1004 11:26:12.212785 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" podStartSLOduration=1.821764828 podStartE2EDuration="2.212708165s" podCreationTimestamp="2025-10-04 11:26:10 +0000 UTC" firstStartedPulling="2025-10-04 11:26:11.249428343 +0000 UTC m=+2148.542079232" lastFinishedPulling="2025-10-04 11:26:11.64037168 +0000 UTC m=+2148.933022569" observedRunningTime="2025-10-04 11:26:12.208450312 +0000 UTC m=+2149.501101201" watchObservedRunningTime="2025-10-04 11:26:12.212708165 +0000 UTC m=+2149.505359064" Oct 04 11:26:12 crc kubenswrapper[4758]: I1004 11:26:12.329931 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:12 crc kubenswrapper[4758]: I1004 11:26:12.331375 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:13 crc kubenswrapper[4758]: I1004 11:26:13.371200 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m8r6s" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="registry-server" probeResult="failure" output=< Oct 04 11:26:13 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:26:13 crc kubenswrapper[4758]: > Oct 04 11:26:22 crc kubenswrapper[4758]: I1004 11:26:22.427993 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:22 crc kubenswrapper[4758]: I1004 11:26:22.495578 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:22 crc kubenswrapper[4758]: I1004 11:26:22.670528 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m8r6s"] Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.341580 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m8r6s" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="registry-server" containerID="cri-o://75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e" gracePeriod=2 Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.816335 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.911986 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-utilities\") pod \"7d805fd7-5d93-4b4f-831e-446e51a318a9\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.912243 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-catalog-content\") pod \"7d805fd7-5d93-4b4f-831e-446e51a318a9\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.912350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5qkf\" (UniqueName: \"kubernetes.io/projected/7d805fd7-5d93-4b4f-831e-446e51a318a9-kube-api-access-w5qkf\") pod \"7d805fd7-5d93-4b4f-831e-446e51a318a9\" (UID: \"7d805fd7-5d93-4b4f-831e-446e51a318a9\") " Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.912810 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-utilities" (OuterVolumeSpecName: "utilities") pod "7d805fd7-5d93-4b4f-831e-446e51a318a9" (UID: "7d805fd7-5d93-4b4f-831e-446e51a318a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:26:24 crc kubenswrapper[4758]: I1004 11:26:24.918591 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d805fd7-5d93-4b4f-831e-446e51a318a9-kube-api-access-w5qkf" (OuterVolumeSpecName: "kube-api-access-w5qkf") pod "7d805fd7-5d93-4b4f-831e-446e51a318a9" (UID: "7d805fd7-5d93-4b4f-831e-446e51a318a9"). InnerVolumeSpecName "kube-api-access-w5qkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.014348 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5qkf\" (UniqueName: \"kubernetes.io/projected/7d805fd7-5d93-4b4f-831e-446e51a318a9-kube-api-access-w5qkf\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.014591 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.015578 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d805fd7-5d93-4b4f-831e-446e51a318a9" (UID: "7d805fd7-5d93-4b4f-831e-446e51a318a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.116095 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d805fd7-5d93-4b4f-831e-446e51a318a9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.352846 4758 generic.go:334] "Generic (PLEG): container finished" podID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerID="75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e" exitCode=0 Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.352911 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerDied","Data":"75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e"} Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.352937 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m8r6s" event={"ID":"7d805fd7-5d93-4b4f-831e-446e51a318a9","Type":"ContainerDied","Data":"98d9ac0ccc1325319d7cced84d059e8aaa4e8c3064de50e0f996b6d87f3a4d35"} Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.352954 4758 scope.go:117] "RemoveContainer" containerID="75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.353013 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m8r6s" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.390639 4758 scope.go:117] "RemoveContainer" containerID="3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.396015 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m8r6s"] Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.410687 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m8r6s"] Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.427340 4758 scope.go:117] "RemoveContainer" containerID="ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.450083 4758 scope.go:117] "RemoveContainer" containerID="75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e" Oct 04 11:26:25 crc kubenswrapper[4758]: E1004 11:26:25.450425 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e\": container with ID starting with 75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e not found: ID does not exist" containerID="75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.450456 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e"} err="failed to get container status \"75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e\": rpc error: code = NotFound desc = could not find container \"75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e\": container with ID starting with 75ac2d246eb12be71567fd6d9b5402d6a16401eddce0d0337922fc625dec1a2e not found: ID does not exist" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.450474 4758 scope.go:117] "RemoveContainer" containerID="3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d" Oct 04 11:26:25 crc kubenswrapper[4758]: E1004 11:26:25.450679 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d\": container with ID starting with 3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d not found: ID does not exist" containerID="3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.450701 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d"} err="failed to get container status \"3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d\": rpc error: code = NotFound desc = could not find container \"3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d\": container with ID starting with 3bbf48654ee8d2263270d4d2fe69c12965de056c46531e4b3c5dbde7b1e34e9d not found: ID does not exist" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.450743 4758 scope.go:117] "RemoveContainer" containerID="ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292" Oct 04 11:26:25 crc kubenswrapper[4758]: E1004 11:26:25.450895 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292\": container with ID starting with ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292 not found: ID does not exist" containerID="ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292" Oct 04 11:26:25 crc kubenswrapper[4758]: I1004 11:26:25.450915 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292"} err="failed to get container status \"ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292\": rpc error: code = NotFound desc = could not find container \"ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292\": container with ID starting with ca895e3d32a7bce73b55a5eafb82460bd869de207f14a12940873e7dc49c9292 not found: ID does not exist" Oct 04 11:26:27 crc kubenswrapper[4758]: I1004 11:26:27.336496 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" path="/var/lib/kubelet/pods/7d805fd7-5d93-4b4f-831e-446e51a318a9/volumes" Oct 04 11:27:31 crc kubenswrapper[4758]: I1004 11:27:31.249583 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:27:31 crc kubenswrapper[4758]: I1004 11:27:31.250422 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:28:01 crc kubenswrapper[4758]: I1004 11:28:01.249461 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:28:01 crc kubenswrapper[4758]: I1004 11:28:01.250044 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.249480 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.250073 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.250166 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.251035 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.251132 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" gracePeriod=600 Oct 04 11:28:31 crc kubenswrapper[4758]: E1004 11:28:31.377585 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.487340 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" exitCode=0 Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.487376 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888"} Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.487418 4758 scope.go:117] "RemoveContainer" containerID="8fb28a3d28b026315f63e9ed9f93326a8b51ed6380ad4d69c7930ab2c772926d" Oct 04 11:28:31 crc kubenswrapper[4758]: I1004 11:28:31.488414 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:28:31 crc kubenswrapper[4758]: E1004 11:28:31.490359 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.694937 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mfj"] Oct 04 11:28:42 crc kubenswrapper[4758]: E1004 11:28:42.696278 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="registry-server" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.696298 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="registry-server" Oct 04 11:28:42 crc kubenswrapper[4758]: E1004 11:28:42.696324 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="extract-content" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.696333 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="extract-content" Oct 04 11:28:42 crc kubenswrapper[4758]: E1004 11:28:42.696353 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="extract-utilities" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.696361 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="extract-utilities" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.696625 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d805fd7-5d93-4b4f-831e-446e51a318a9" containerName="registry-server" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.699484 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.711744 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mfj"] Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.786828 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-catalog-content\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.786948 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-utilities\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.787006 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkbrt\" (UniqueName: \"kubernetes.io/projected/cc4f4974-162b-4793-b254-1ec1a2e3db3e-kube-api-access-dkbrt\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.888669 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-catalog-content\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.889140 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-utilities\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.889212 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkbrt\" (UniqueName: \"kubernetes.io/projected/cc4f4974-162b-4793-b254-1ec1a2e3db3e-kube-api-access-dkbrt\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.889596 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-catalog-content\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.889846 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-utilities\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:42 crc kubenswrapper[4758]: I1004 11:28:42.907957 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkbrt\" (UniqueName: \"kubernetes.io/projected/cc4f4974-162b-4793-b254-1ec1a2e3db3e-kube-api-access-dkbrt\") pod \"redhat-marketplace-c6mfj\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:43 crc kubenswrapper[4758]: I1004 11:28:43.042477 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:43 crc kubenswrapper[4758]: I1004 11:28:43.350919 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mfj"] Oct 04 11:28:43 crc kubenswrapper[4758]: I1004 11:28:43.598563 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerStarted","Data":"f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177"} Oct 04 11:28:43 crc kubenswrapper[4758]: I1004 11:28:43.598599 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerStarted","Data":"66982921d9a57c92b3f12d767609cfe05a7b493cbed67f543edacf53e565bfbd"} Oct 04 11:28:44 crc kubenswrapper[4758]: I1004 11:28:44.607777 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerID="f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177" exitCode=0 Oct 04 11:28:44 crc kubenswrapper[4758]: I1004 11:28:44.607834 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerDied","Data":"f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177"} Oct 04 11:28:44 crc kubenswrapper[4758]: I1004 11:28:44.612278 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:28:46 crc kubenswrapper[4758]: I1004 11:28:46.326090 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:28:46 crc kubenswrapper[4758]: E1004 11:28:46.326646 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:28:46 crc kubenswrapper[4758]: I1004 11:28:46.634619 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerStarted","Data":"0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26"} Oct 04 11:28:47 crc kubenswrapper[4758]: I1004 11:28:47.643230 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerID="0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26" exitCode=0 Oct 04 11:28:47 crc kubenswrapper[4758]: I1004 11:28:47.643289 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerDied","Data":"0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26"} Oct 04 11:28:48 crc kubenswrapper[4758]: I1004 11:28:48.652687 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerStarted","Data":"63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8"} Oct 04 11:28:48 crc kubenswrapper[4758]: I1004 11:28:48.678471 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c6mfj" podStartSLOduration=3.219585202 podStartE2EDuration="6.678451315s" podCreationTimestamp="2025-10-04 11:28:42 +0000 UTC" firstStartedPulling="2025-10-04 11:28:44.611923426 +0000 UTC m=+2301.904574315" lastFinishedPulling="2025-10-04 11:28:48.070789539 +0000 UTC m=+2305.363440428" observedRunningTime="2025-10-04 11:28:48.671965902 +0000 UTC m=+2305.964616791" watchObservedRunningTime="2025-10-04 11:28:48.678451315 +0000 UTC m=+2305.971102204" Oct 04 11:28:53 crc kubenswrapper[4758]: I1004 11:28:53.048037 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:53 crc kubenswrapper[4758]: I1004 11:28:53.048650 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:53 crc kubenswrapper[4758]: I1004 11:28:53.108976 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:53 crc kubenswrapper[4758]: I1004 11:28:53.766566 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:53 crc kubenswrapper[4758]: I1004 11:28:53.835769 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mfj"] Oct 04 11:28:55 crc kubenswrapper[4758]: I1004 11:28:55.719600 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c6mfj" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="registry-server" containerID="cri-o://63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8" gracePeriod=2 Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.207455 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.342675 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-utilities\") pod \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.342799 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-catalog-content\") pod \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.342848 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkbrt\" (UniqueName: \"kubernetes.io/projected/cc4f4974-162b-4793-b254-1ec1a2e3db3e-kube-api-access-dkbrt\") pod \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\" (UID: \"cc4f4974-162b-4793-b254-1ec1a2e3db3e\") " Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.344840 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-utilities" (OuterVolumeSpecName: "utilities") pod "cc4f4974-162b-4793-b254-1ec1a2e3db3e" (UID: "cc4f4974-162b-4793-b254-1ec1a2e3db3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.351877 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc4f4974-162b-4793-b254-1ec1a2e3db3e-kube-api-access-dkbrt" (OuterVolumeSpecName: "kube-api-access-dkbrt") pod "cc4f4974-162b-4793-b254-1ec1a2e3db3e" (UID: "cc4f4974-162b-4793-b254-1ec1a2e3db3e"). InnerVolumeSpecName "kube-api-access-dkbrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.359019 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc4f4974-162b-4793-b254-1ec1a2e3db3e" (UID: "cc4f4974-162b-4793-b254-1ec1a2e3db3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.445296 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.445326 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkbrt\" (UniqueName: \"kubernetes.io/projected/cc4f4974-162b-4793-b254-1ec1a2e3db3e-kube-api-access-dkbrt\") on node \"crc\" DevicePath \"\"" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.445336 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc4f4974-162b-4793-b254-1ec1a2e3db3e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.731347 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerID="63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8" exitCode=0 Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.731398 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerDied","Data":"63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8"} Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.731462 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mfj" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.733292 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mfj" event={"ID":"cc4f4974-162b-4793-b254-1ec1a2e3db3e","Type":"ContainerDied","Data":"66982921d9a57c92b3f12d767609cfe05a7b493cbed67f543edacf53e565bfbd"} Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.733432 4758 scope.go:117] "RemoveContainer" containerID="63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.767654 4758 scope.go:117] "RemoveContainer" containerID="0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.784906 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mfj"] Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.801503 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mfj"] Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.817949 4758 scope.go:117] "RemoveContainer" containerID="f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.841780 4758 scope.go:117] "RemoveContainer" containerID="63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8" Oct 04 11:28:56 crc kubenswrapper[4758]: E1004 11:28:56.842225 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8\": container with ID starting with 63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8 not found: ID does not exist" containerID="63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.842256 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8"} err="failed to get container status \"63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8\": rpc error: code = NotFound desc = could not find container \"63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8\": container with ID starting with 63d8055148efc1c12ab9ece87f2fc6770571c93a2fdf5a48a2adae0c291320c8 not found: ID does not exist" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.842282 4758 scope.go:117] "RemoveContainer" containerID="0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26" Oct 04 11:28:56 crc kubenswrapper[4758]: E1004 11:28:56.842547 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26\": container with ID starting with 0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26 not found: ID does not exist" containerID="0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.842570 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26"} err="failed to get container status \"0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26\": rpc error: code = NotFound desc = could not find container \"0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26\": container with ID starting with 0cf39561ef9a8c55146070ca1713311ff69f6edb2948bc7bf4f19db7c2402d26 not found: ID does not exist" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.842582 4758 scope.go:117] "RemoveContainer" containerID="f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177" Oct 04 11:28:56 crc kubenswrapper[4758]: E1004 11:28:56.842945 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177\": container with ID starting with f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177 not found: ID does not exist" containerID="f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177" Oct 04 11:28:56 crc kubenswrapper[4758]: I1004 11:28:56.842968 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177"} err="failed to get container status \"f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177\": rpc error: code = NotFound desc = could not find container \"f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177\": container with ID starting with f65bc2dcd34bcfee192c96acb797d2d19b1918e1f56192564401e2be62e00177 not found: ID does not exist" Oct 04 11:28:57 crc kubenswrapper[4758]: I1004 11:28:57.335137 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" path="/var/lib/kubelet/pods/cc4f4974-162b-4793-b254-1ec1a2e3db3e/volumes" Oct 04 11:28:58 crc kubenswrapper[4758]: I1004 11:28:58.327514 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:28:58 crc kubenswrapper[4758]: E1004 11:28:58.328416 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:29:10 crc kubenswrapper[4758]: I1004 11:29:10.326396 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:29:10 crc kubenswrapper[4758]: E1004 11:29:10.327318 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:29:23 crc kubenswrapper[4758]: I1004 11:29:23.330757 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:29:23 crc kubenswrapper[4758]: E1004 11:29:23.332673 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:29:37 crc kubenswrapper[4758]: I1004 11:29:37.326535 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:29:37 crc kubenswrapper[4758]: E1004 11:29:37.327374 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:29:52 crc kubenswrapper[4758]: I1004 11:29:52.326694 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:29:52 crc kubenswrapper[4758]: E1004 11:29:52.327808 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.154009 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v"] Oct 04 11:30:00 crc kubenswrapper[4758]: E1004 11:30:00.155230 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.155253 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[4758]: E1004 11:30:00.155318 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="extract-utilities" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.155329 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="extract-utilities" Oct 04 11:30:00 crc kubenswrapper[4758]: E1004 11:30:00.155350 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="extract-content" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.155360 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="extract-content" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.155645 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc4f4974-162b-4793-b254-1ec1a2e3db3e" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.156676 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.166077 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.166296 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.166804 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v"] Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.305155 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zdq\" (UniqueName: \"kubernetes.io/projected/2132fe7d-21cb-4aa5-abd5-ac438011c032-kube-api-access-d9zdq\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.305987 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2132fe7d-21cb-4aa5-abd5-ac438011c032-config-volume\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.306524 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2132fe7d-21cb-4aa5-abd5-ac438011c032-secret-volume\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.408419 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2132fe7d-21cb-4aa5-abd5-ac438011c032-secret-volume\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.408519 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9zdq\" (UniqueName: \"kubernetes.io/projected/2132fe7d-21cb-4aa5-abd5-ac438011c032-kube-api-access-d9zdq\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.408638 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2132fe7d-21cb-4aa5-abd5-ac438011c032-config-volume\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.409535 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2132fe7d-21cb-4aa5-abd5-ac438011c032-config-volume\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.413906 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2132fe7d-21cb-4aa5-abd5-ac438011c032-secret-volume\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.427357 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9zdq\" (UniqueName: \"kubernetes.io/projected/2132fe7d-21cb-4aa5-abd5-ac438011c032-kube-api-access-d9zdq\") pod \"collect-profiles-29326290-m2m4v\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.477372 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:00 crc kubenswrapper[4758]: I1004 11:30:00.935299 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v"] Oct 04 11:30:01 crc kubenswrapper[4758]: I1004 11:30:01.345650 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" event={"ID":"2132fe7d-21cb-4aa5-abd5-ac438011c032","Type":"ContainerStarted","Data":"5e123aeb016b9d846db7d76a560f15e22e9c64a6c034db45378416ef795dd7dc"} Oct 04 11:30:01 crc kubenswrapper[4758]: I1004 11:30:01.345693 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" event={"ID":"2132fe7d-21cb-4aa5-abd5-ac438011c032","Type":"ContainerStarted","Data":"1817cf4ddb633600e7cb2bf2dd932eda784c00bcd91c18ddc401bcb5118590f5"} Oct 04 11:30:01 crc kubenswrapper[4758]: I1004 11:30:01.371045 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" podStartSLOduration=1.371030208 podStartE2EDuration="1.371030208s" podCreationTimestamp="2025-10-04 11:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:30:01.365204063 +0000 UTC m=+2378.657854972" watchObservedRunningTime="2025-10-04 11:30:01.371030208 +0000 UTC m=+2378.663681097" Oct 04 11:30:02 crc kubenswrapper[4758]: I1004 11:30:02.357501 4758 generic.go:334] "Generic (PLEG): container finished" podID="2132fe7d-21cb-4aa5-abd5-ac438011c032" containerID="5e123aeb016b9d846db7d76a560f15e22e9c64a6c034db45378416ef795dd7dc" exitCode=0 Oct 04 11:30:02 crc kubenswrapper[4758]: I1004 11:30:02.357632 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" event={"ID":"2132fe7d-21cb-4aa5-abd5-ac438011c032","Type":"ContainerDied","Data":"5e123aeb016b9d846db7d76a560f15e22e9c64a6c034db45378416ef795dd7dc"} Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.675552 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.815871 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9zdq\" (UniqueName: \"kubernetes.io/projected/2132fe7d-21cb-4aa5-abd5-ac438011c032-kube-api-access-d9zdq\") pod \"2132fe7d-21cb-4aa5-abd5-ac438011c032\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.816174 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2132fe7d-21cb-4aa5-abd5-ac438011c032-config-volume\") pod \"2132fe7d-21cb-4aa5-abd5-ac438011c032\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.816228 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2132fe7d-21cb-4aa5-abd5-ac438011c032-secret-volume\") pod \"2132fe7d-21cb-4aa5-abd5-ac438011c032\" (UID: \"2132fe7d-21cb-4aa5-abd5-ac438011c032\") " Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.816724 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2132fe7d-21cb-4aa5-abd5-ac438011c032-config-volume" (OuterVolumeSpecName: "config-volume") pod "2132fe7d-21cb-4aa5-abd5-ac438011c032" (UID: "2132fe7d-21cb-4aa5-abd5-ac438011c032"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.821897 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2132fe7d-21cb-4aa5-abd5-ac438011c032-kube-api-access-d9zdq" (OuterVolumeSpecName: "kube-api-access-d9zdq") pod "2132fe7d-21cb-4aa5-abd5-ac438011c032" (UID: "2132fe7d-21cb-4aa5-abd5-ac438011c032"). InnerVolumeSpecName "kube-api-access-d9zdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.822010 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2132fe7d-21cb-4aa5-abd5-ac438011c032-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2132fe7d-21cb-4aa5-abd5-ac438011c032" (UID: "2132fe7d-21cb-4aa5-abd5-ac438011c032"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.918393 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2132fe7d-21cb-4aa5-abd5-ac438011c032-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.918433 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2132fe7d-21cb-4aa5-abd5-ac438011c032-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:03 crc kubenswrapper[4758]: I1004 11:30:03.918448 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9zdq\" (UniqueName: \"kubernetes.io/projected/2132fe7d-21cb-4aa5-abd5-ac438011c032-kube-api-access-d9zdq\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:04 crc kubenswrapper[4758]: I1004 11:30:04.325270 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:30:04 crc kubenswrapper[4758]: E1004 11:30:04.325504 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:30:04 crc kubenswrapper[4758]: I1004 11:30:04.375004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" event={"ID":"2132fe7d-21cb-4aa5-abd5-ac438011c032","Type":"ContainerDied","Data":"1817cf4ddb633600e7cb2bf2dd932eda784c00bcd91c18ddc401bcb5118590f5"} Oct 04 11:30:04 crc kubenswrapper[4758]: I1004 11:30:04.375048 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-m2m4v" Oct 04 11:30:04 crc kubenswrapper[4758]: I1004 11:30:04.375055 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1817cf4ddb633600e7cb2bf2dd932eda784c00bcd91c18ddc401bcb5118590f5" Oct 04 11:30:04 crc kubenswrapper[4758]: I1004 11:30:04.433993 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm"] Oct 04 11:30:04 crc kubenswrapper[4758]: I1004 11:30:04.440520 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-p5jpm"] Oct 04 11:30:05 crc kubenswrapper[4758]: I1004 11:30:05.337473 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7c28209-920e-49c3-9b2b-334dc06df888" path="/var/lib/kubelet/pods/a7c28209-920e-49c3-9b2b-334dc06df888/volumes" Oct 04 11:30:17 crc kubenswrapper[4758]: I1004 11:30:17.326672 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:30:17 crc kubenswrapper[4758]: E1004 11:30:17.327563 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:30:28 crc kubenswrapper[4758]: I1004 11:30:28.327079 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:30:28 crc kubenswrapper[4758]: E1004 11:30:28.327944 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:30:42 crc kubenswrapper[4758]: I1004 11:30:42.326147 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:30:42 crc kubenswrapper[4758]: E1004 11:30:42.327229 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:30:55 crc kubenswrapper[4758]: I1004 11:30:55.326062 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:30:55 crc kubenswrapper[4758]: E1004 11:30:55.327057 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:30:55 crc kubenswrapper[4758]: I1004 11:30:55.849312 4758 generic.go:334] "Generic (PLEG): container finished" podID="6ec33929-038a-47ec-9240-42f88c26168e" containerID="28c85d8ef1e3813c224ecd2eb8f60be39043a06a89185ab7e33186bc9514b416" exitCode=0 Oct 04 11:30:55 crc kubenswrapper[4758]: I1004 11:30:55.849373 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" event={"ID":"6ec33929-038a-47ec-9240-42f88c26168e","Type":"ContainerDied","Data":"28c85d8ef1e3813c224ecd2eb8f60be39043a06a89185ab7e33186bc9514b416"} Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.365242 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.488059 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-secret-0\") pod \"6ec33929-038a-47ec-9240-42f88c26168e\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.488209 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-combined-ca-bundle\") pod \"6ec33929-038a-47ec-9240-42f88c26168e\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.488244 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-inventory\") pod \"6ec33929-038a-47ec-9240-42f88c26168e\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.488276 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5d2v\" (UniqueName: \"kubernetes.io/projected/6ec33929-038a-47ec-9240-42f88c26168e-kube-api-access-p5d2v\") pod \"6ec33929-038a-47ec-9240-42f88c26168e\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.488504 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-ssh-key\") pod \"6ec33929-038a-47ec-9240-42f88c26168e\" (UID: \"6ec33929-038a-47ec-9240-42f88c26168e\") " Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.493885 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ec33929-038a-47ec-9240-42f88c26168e-kube-api-access-p5d2v" (OuterVolumeSpecName: "kube-api-access-p5d2v") pod "6ec33929-038a-47ec-9240-42f88c26168e" (UID: "6ec33929-038a-47ec-9240-42f88c26168e"). InnerVolumeSpecName "kube-api-access-p5d2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.498295 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6ec33929-038a-47ec-9240-42f88c26168e" (UID: "6ec33929-038a-47ec-9240-42f88c26168e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.516841 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "6ec33929-038a-47ec-9240-42f88c26168e" (UID: "6ec33929-038a-47ec-9240-42f88c26168e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.539303 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-inventory" (OuterVolumeSpecName: "inventory") pod "6ec33929-038a-47ec-9240-42f88c26168e" (UID: "6ec33929-038a-47ec-9240-42f88c26168e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.540887 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ec33929-038a-47ec-9240-42f88c26168e" (UID: "6ec33929-038a-47ec-9240-42f88c26168e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.590997 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.591185 4758 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.591247 4758 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.591378 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ec33929-038a-47ec-9240-42f88c26168e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.591434 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5d2v\" (UniqueName: \"kubernetes.io/projected/6ec33929-038a-47ec-9240-42f88c26168e-kube-api-access-p5d2v\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.881200 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" event={"ID":"6ec33929-038a-47ec-9240-42f88c26168e","Type":"ContainerDied","Data":"9f13545d1c8ac3bc88d595dec94526881c77fed0993e910c07fe4c6f24260904"} Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.881242 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f13545d1c8ac3bc88d595dec94526881c77fed0993e910c07fe4c6f24260904" Oct 04 11:30:57 crc kubenswrapper[4758]: I1004 11:30:57.881318 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m97jl" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.042544 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7"] Oct 04 11:30:58 crc kubenswrapper[4758]: E1004 11:30:58.042898 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec33929-038a-47ec-9240-42f88c26168e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.042917 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec33929-038a-47ec-9240-42f88c26168e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 11:30:58 crc kubenswrapper[4758]: E1004 11:30:58.042942 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2132fe7d-21cb-4aa5-abd5-ac438011c032" containerName="collect-profiles" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.042951 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2132fe7d-21cb-4aa5-abd5-ac438011c032" containerName="collect-profiles" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.043120 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ec33929-038a-47ec-9240-42f88c26168e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.043142 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2132fe7d-21cb-4aa5-abd5-ac438011c032" containerName="collect-profiles" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.048161 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.051712 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.051798 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.051862 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.051867 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.051935 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.051967 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.057137 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.062627 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7"] Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201503 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201561 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/02009f56-9d0a-404f-8f19-475ac09aad43-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201585 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201607 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201650 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201688 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201706 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201742 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.201777 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8krh\" (UniqueName: \"kubernetes.io/projected/02009f56-9d0a-404f-8f19-475ac09aad43-kube-api-access-l8krh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.303444 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.303670 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.303766 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.303864 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8krh\" (UniqueName: \"kubernetes.io/projected/02009f56-9d0a-404f-8f19-475ac09aad43-kube-api-access-l8krh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.303969 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.304054 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/02009f56-9d0a-404f-8f19-475ac09aad43-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.304135 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.304217 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.304316 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.305492 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/02009f56-9d0a-404f-8f19-475ac09aad43-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.308367 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.309055 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.310024 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.310496 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.310540 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.310709 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.312568 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.326450 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8krh\" (UniqueName: \"kubernetes.io/projected/02009f56-9d0a-404f-8f19-475ac09aad43-kube-api-access-l8krh\") pod \"nova-edpm-deployment-openstack-edpm-ipam-cw4t7\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.377204 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.434526 4758 scope.go:117] "RemoveContainer" containerID="cb4935e21eae4fc56d4436b2fedb3882cb8b772254b3e8e01904491898c7e8df" Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.731390 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7"] Oct 04 11:30:58 crc kubenswrapper[4758]: I1004 11:30:58.894242 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" event={"ID":"02009f56-9d0a-404f-8f19-475ac09aad43","Type":"ContainerStarted","Data":"3e6ec121f0b316bd8ab08b79fb08ff27f9a9878818cc8d41bbad61835a28844c"} Oct 04 11:30:59 crc kubenswrapper[4758]: I1004 11:30:59.905326 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" event={"ID":"02009f56-9d0a-404f-8f19-475ac09aad43","Type":"ContainerStarted","Data":"50579d040c7a096800ced8e32c5d0fa9464f5bc5afdcb09d28a08fbfa66e37bb"} Oct 04 11:30:59 crc kubenswrapper[4758]: I1004 11:30:59.938261 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" podStartSLOduration=1.490092218 podStartE2EDuration="1.938235677s" podCreationTimestamp="2025-10-04 11:30:58 +0000 UTC" firstStartedPulling="2025-10-04 11:30:58.732584207 +0000 UTC m=+2436.025235116" lastFinishedPulling="2025-10-04 11:30:59.180727676 +0000 UTC m=+2436.473378575" observedRunningTime="2025-10-04 11:30:59.92708752 +0000 UTC m=+2437.219738419" watchObservedRunningTime="2025-10-04 11:30:59.938235677 +0000 UTC m=+2437.230886566" Oct 04 11:31:10 crc kubenswrapper[4758]: I1004 11:31:10.326689 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:31:10 crc kubenswrapper[4758]: E1004 11:31:10.327459 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:31:24 crc kubenswrapper[4758]: I1004 11:31:24.325923 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:31:24 crc kubenswrapper[4758]: E1004 11:31:24.326783 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:31:37 crc kubenswrapper[4758]: I1004 11:31:37.326913 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:31:37 crc kubenswrapper[4758]: E1004 11:31:37.327787 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:31:52 crc kubenswrapper[4758]: I1004 11:31:52.326291 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:31:52 crc kubenswrapper[4758]: E1004 11:31:52.327043 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:32:03 crc kubenswrapper[4758]: I1004 11:32:03.332738 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:32:03 crc kubenswrapper[4758]: E1004 11:32:03.333606 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:32:18 crc kubenswrapper[4758]: I1004 11:32:18.325375 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:32:18 crc kubenswrapper[4758]: E1004 11:32:18.326038 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:32:31 crc kubenswrapper[4758]: I1004 11:32:31.325395 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:32:31 crc kubenswrapper[4758]: E1004 11:32:31.326139 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:32:44 crc kubenswrapper[4758]: I1004 11:32:44.326487 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:32:44 crc kubenswrapper[4758]: E1004 11:32:44.327411 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:32:57 crc kubenswrapper[4758]: I1004 11:32:57.326082 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:32:57 crc kubenswrapper[4758]: E1004 11:32:57.328091 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:33:08 crc kubenswrapper[4758]: I1004 11:33:08.325613 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:33:08 crc kubenswrapper[4758]: E1004 11:33:08.326282 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:33:19 crc kubenswrapper[4758]: I1004 11:33:19.325487 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:33:19 crc kubenswrapper[4758]: E1004 11:33:19.327268 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:33:31 crc kubenswrapper[4758]: I1004 11:33:31.326139 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:33:32 crc kubenswrapper[4758]: I1004 11:33:32.373879 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"86475db01c8dfb4eef260997889015747080409f7c657306fa9c606433685e87"} Oct 04 11:34:54 crc kubenswrapper[4758]: I1004 11:34:54.085674 4758 generic.go:334] "Generic (PLEG): container finished" podID="02009f56-9d0a-404f-8f19-475ac09aad43" containerID="50579d040c7a096800ced8e32c5d0fa9464f5bc5afdcb09d28a08fbfa66e37bb" exitCode=0 Oct 04 11:34:54 crc kubenswrapper[4758]: I1004 11:34:54.085755 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" event={"ID":"02009f56-9d0a-404f-8f19-475ac09aad43","Type":"ContainerDied","Data":"50579d040c7a096800ced8e32c5d0fa9464f5bc5afdcb09d28a08fbfa66e37bb"} Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.542219 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.705802 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-combined-ca-bundle\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.705859 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/02009f56-9d0a-404f-8f19-475ac09aad43-nova-extra-config-0\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706030 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-1\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706071 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-0\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706122 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-0\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706161 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-1\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706186 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8krh\" (UniqueName: \"kubernetes.io/projected/02009f56-9d0a-404f-8f19-475ac09aad43-kube-api-access-l8krh\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706213 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-ssh-key\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.706241 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-inventory\") pod \"02009f56-9d0a-404f-8f19-475ac09aad43\" (UID: \"02009f56-9d0a-404f-8f19-475ac09aad43\") " Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.713565 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.714267 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02009f56-9d0a-404f-8f19-475ac09aad43-kube-api-access-l8krh" (OuterVolumeSpecName: "kube-api-access-l8krh") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "kube-api-access-l8krh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.736912 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.736958 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02009f56-9d0a-404f-8f19-475ac09aad43-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.751676 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-inventory" (OuterVolumeSpecName: "inventory") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.752234 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.757838 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.760607 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.768797 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "02009f56-9d0a-404f-8f19-475ac09aad43" (UID: "02009f56-9d0a-404f-8f19-475ac09aad43"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808782 4758 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/02009f56-9d0a-404f-8f19-475ac09aad43-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808806 4758 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808816 4758 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808825 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808833 4758 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808842 4758 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808849 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8krh\" (UniqueName: \"kubernetes.io/projected/02009f56-9d0a-404f-8f19-475ac09aad43-kube-api-access-l8krh\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808859 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:55 crc kubenswrapper[4758]: I1004 11:34:55.808867 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02009f56-9d0a-404f-8f19-475ac09aad43-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.109854 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" event={"ID":"02009f56-9d0a-404f-8f19-475ac09aad43","Type":"ContainerDied","Data":"3e6ec121f0b316bd8ab08b79fb08ff27f9a9878818cc8d41bbad61835a28844c"} Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.109896 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e6ec121f0b316bd8ab08b79fb08ff27f9a9878818cc8d41bbad61835a28844c" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.109979 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-cw4t7" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.238816 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs"] Oct 04 11:34:56 crc kubenswrapper[4758]: E1004 11:34:56.239486 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02009f56-9d0a-404f-8f19-475ac09aad43" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.239504 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="02009f56-9d0a-404f-8f19-475ac09aad43" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.239681 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="02009f56-9d0a-404f-8f19-475ac09aad43" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.240295 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.241804 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.242611 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.244921 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.244970 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-lqz7s" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.245089 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.258087 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs"] Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426732 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w4ml\" (UniqueName: \"kubernetes.io/projected/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-kube-api-access-7w4ml\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426775 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426870 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426895 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426947 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-3\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-3\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.426979 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.427032 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-4\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.427086 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529016 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-4\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529144 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529177 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w4ml\" (UniqueName: \"kubernetes.io/projected/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-kube-api-access-7w4ml\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529202 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529227 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529286 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529315 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529369 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-3\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-3\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.529392 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.533771 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.533799 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.535675 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.536122 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.537032 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.539272 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-4\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-4\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.539970 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.543208 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-3\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-3\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.564447 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w4ml\" (UniqueName: \"kubernetes.io/projected/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-kube-api-access-7w4ml\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:56 crc kubenswrapper[4758]: I1004 11:34:56.856220 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:34:57 crc kubenswrapper[4758]: I1004 11:34:57.390307 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs"] Oct 04 11:34:57 crc kubenswrapper[4758]: I1004 11:34:57.401033 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:34:58 crc kubenswrapper[4758]: I1004 11:34:58.126359 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" event={"ID":"acd41b88-adbc-46b6-b2f0-dde72b71dbf5","Type":"ContainerStarted","Data":"c3839868e774c22a954803fb030f62032149d6893b79b84a322cbb2be18adc22"} Oct 04 11:34:59 crc kubenswrapper[4758]: I1004 11:34:59.139313 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" event={"ID":"acd41b88-adbc-46b6-b2f0-dde72b71dbf5","Type":"ContainerStarted","Data":"abe8b5ab1106b7089cbfa7c06e70acfff41c77bba6f2d80d594957df58c2457a"} Oct 04 11:34:59 crc kubenswrapper[4758]: I1004 11:34:59.167778 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" podStartSLOduration=2.680439658 podStartE2EDuration="3.167756341s" podCreationTimestamp="2025-10-04 11:34:56 +0000 UTC" firstStartedPulling="2025-10-04 11:34:57.400766116 +0000 UTC m=+2674.693417005" lastFinishedPulling="2025-10-04 11:34:57.888082799 +0000 UTC m=+2675.180733688" observedRunningTime="2025-10-04 11:34:59.158857424 +0000 UTC m=+2676.451508333" watchObservedRunningTime="2025-10-04 11:34:59.167756341 +0000 UTC m=+2676.460407240" Oct 04 11:35:31 crc kubenswrapper[4758]: I1004 11:35:31.249174 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:35:31 crc kubenswrapper[4758]: I1004 11:35:31.249746 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:36:01 crc kubenswrapper[4758]: I1004 11:36:01.249545 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:36:01 crc kubenswrapper[4758]: I1004 11:36:01.250864 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.514710 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5twc"] Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.517355 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.542244 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5twc"] Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.568787 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-utilities\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.568926 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-catalog-content\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.568995 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbz9d\" (UniqueName: \"kubernetes.io/projected/d92e5692-ff15-4fc2-98e6-fefdb083f629-kube-api-access-xbz9d\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.671395 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-catalog-content\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.671474 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbz9d\" (UniqueName: \"kubernetes.io/projected/d92e5692-ff15-4fc2-98e6-fefdb083f629-kube-api-access-xbz9d\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.671562 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-utilities\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.672080 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-utilities\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.672259 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-catalog-content\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.691890 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbz9d\" (UniqueName: \"kubernetes.io/projected/d92e5692-ff15-4fc2-98e6-fefdb083f629-kube-api-access-xbz9d\") pod \"redhat-operators-r5twc\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:17 crc kubenswrapper[4758]: I1004 11:36:17.837055 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:18 crc kubenswrapper[4758]: I1004 11:36:18.301002 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5twc"] Oct 04 11:36:18 crc kubenswrapper[4758]: I1004 11:36:18.873520 4758 generic.go:334] "Generic (PLEG): container finished" podID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerID="eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60" exitCode=0 Oct 04 11:36:18 crc kubenswrapper[4758]: I1004 11:36:18.873628 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5twc" event={"ID":"d92e5692-ff15-4fc2-98e6-fefdb083f629","Type":"ContainerDied","Data":"eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60"} Oct 04 11:36:18 crc kubenswrapper[4758]: I1004 11:36:18.873861 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5twc" event={"ID":"d92e5692-ff15-4fc2-98e6-fefdb083f629","Type":"ContainerStarted","Data":"eb6cc96a3513a7f1df2a6c1a750904b167d88b502115edfcfcbede9ca76a1afd"} Oct 04 11:36:20 crc kubenswrapper[4758]: I1004 11:36:20.898466 4758 generic.go:334] "Generic (PLEG): container finished" podID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerID="687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4" exitCode=0 Oct 04 11:36:20 crc kubenswrapper[4758]: I1004 11:36:20.898528 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5twc" event={"ID":"d92e5692-ff15-4fc2-98e6-fefdb083f629","Type":"ContainerDied","Data":"687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4"} Oct 04 11:36:21 crc kubenswrapper[4758]: I1004 11:36:21.876843 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xcf4"] Oct 04 11:36:21 crc kubenswrapper[4758]: I1004 11:36:21.879091 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:21 crc kubenswrapper[4758]: I1004 11:36:21.889475 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xcf4"] Oct 04 11:36:21 crc kubenswrapper[4758]: I1004 11:36:21.919306 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5twc" event={"ID":"d92e5692-ff15-4fc2-98e6-fefdb083f629","Type":"ContainerStarted","Data":"d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625"} Oct 04 11:36:21 crc kubenswrapper[4758]: I1004 11:36:21.939453 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5twc" podStartSLOduration=2.456746829 podStartE2EDuration="4.939434713s" podCreationTimestamp="2025-10-04 11:36:17 +0000 UTC" firstStartedPulling="2025-10-04 11:36:18.875184926 +0000 UTC m=+2756.167835825" lastFinishedPulling="2025-10-04 11:36:21.35787279 +0000 UTC m=+2758.650523709" observedRunningTime="2025-10-04 11:36:21.933653749 +0000 UTC m=+2759.226304638" watchObservedRunningTime="2025-10-04 11:36:21.939434713 +0000 UTC m=+2759.232085602" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.057607 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srdzj\" (UniqueName: \"kubernetes.io/projected/9a6b6bdc-dca8-42b8-bf75-973f402905e9-kube-api-access-srdzj\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.057673 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-catalog-content\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.057703 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-utilities\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.159416 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srdzj\" (UniqueName: \"kubernetes.io/projected/9a6b6bdc-dca8-42b8-bf75-973f402905e9-kube-api-access-srdzj\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.159485 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-catalog-content\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.159514 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-utilities\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.160127 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-utilities\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.160471 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-catalog-content\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.193199 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srdzj\" (UniqueName: \"kubernetes.io/projected/9a6b6bdc-dca8-42b8-bf75-973f402905e9-kube-api-access-srdzj\") pod \"certified-operators-6xcf4\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.202076 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.827413 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xcf4"] Oct 04 11:36:22 crc kubenswrapper[4758]: I1004 11:36:22.933674 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerStarted","Data":"55d314f7c0ad568f97e079b786888efc68a675dd319acdb007ef339d446cb979"} Oct 04 11:36:23 crc kubenswrapper[4758]: I1004 11:36:23.943311 4758 generic.go:334] "Generic (PLEG): container finished" podID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerID="62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54" exitCode=0 Oct 04 11:36:23 crc kubenswrapper[4758]: I1004 11:36:23.944373 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerDied","Data":"62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54"} Oct 04 11:36:24 crc kubenswrapper[4758]: I1004 11:36:24.954548 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerStarted","Data":"aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4"} Oct 04 11:36:26 crc kubenswrapper[4758]: I1004 11:36:26.974811 4758 generic.go:334] "Generic (PLEG): container finished" podID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerID="aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4" exitCode=0 Oct 04 11:36:26 crc kubenswrapper[4758]: I1004 11:36:26.974930 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerDied","Data":"aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4"} Oct 04 11:36:27 crc kubenswrapper[4758]: I1004 11:36:27.838191 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:27 crc kubenswrapper[4758]: I1004 11:36:27.838537 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:27 crc kubenswrapper[4758]: I1004 11:36:27.901123 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:27 crc kubenswrapper[4758]: I1004 11:36:27.984429 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerStarted","Data":"3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78"} Oct 04 11:36:28 crc kubenswrapper[4758]: I1004 11:36:28.006757 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xcf4" podStartSLOduration=3.5058468879999998 podStartE2EDuration="7.006735918s" podCreationTimestamp="2025-10-04 11:36:21 +0000 UTC" firstStartedPulling="2025-10-04 11:36:23.946318411 +0000 UTC m=+2761.238969300" lastFinishedPulling="2025-10-04 11:36:27.447207431 +0000 UTC m=+2764.739858330" observedRunningTime="2025-10-04 11:36:27.999370792 +0000 UTC m=+2765.292021681" watchObservedRunningTime="2025-10-04 11:36:28.006735918 +0000 UTC m=+2765.299386817" Oct 04 11:36:28 crc kubenswrapper[4758]: I1004 11:36:28.034809 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.075729 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5twc"] Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.076677 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r5twc" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="registry-server" containerID="cri-o://d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625" gracePeriod=2 Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.541305 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.635695 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbz9d\" (UniqueName: \"kubernetes.io/projected/d92e5692-ff15-4fc2-98e6-fefdb083f629-kube-api-access-xbz9d\") pod \"d92e5692-ff15-4fc2-98e6-fefdb083f629\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.635859 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-utilities\") pod \"d92e5692-ff15-4fc2-98e6-fefdb083f629\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.635897 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-catalog-content\") pod \"d92e5692-ff15-4fc2-98e6-fefdb083f629\" (UID: \"d92e5692-ff15-4fc2-98e6-fefdb083f629\") " Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.638861 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-utilities" (OuterVolumeSpecName: "utilities") pod "d92e5692-ff15-4fc2-98e6-fefdb083f629" (UID: "d92e5692-ff15-4fc2-98e6-fefdb083f629"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.642290 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d92e5692-ff15-4fc2-98e6-fefdb083f629-kube-api-access-xbz9d" (OuterVolumeSpecName: "kube-api-access-xbz9d") pod "d92e5692-ff15-4fc2-98e6-fefdb083f629" (UID: "d92e5692-ff15-4fc2-98e6-fefdb083f629"). InnerVolumeSpecName "kube-api-access-xbz9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.737685 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbz9d\" (UniqueName: \"kubernetes.io/projected/d92e5692-ff15-4fc2-98e6-fefdb083f629-kube-api-access-xbz9d\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.737725 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.746635 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d92e5692-ff15-4fc2-98e6-fefdb083f629" (UID: "d92e5692-ff15-4fc2-98e6-fefdb083f629"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:36:30 crc kubenswrapper[4758]: I1004 11:36:30.839125 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d92e5692-ff15-4fc2-98e6-fefdb083f629-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.014798 4758 generic.go:334] "Generic (PLEG): container finished" podID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerID="d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625" exitCode=0 Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.014850 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5twc" event={"ID":"d92e5692-ff15-4fc2-98e6-fefdb083f629","Type":"ContainerDied","Data":"d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625"} Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.014886 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5twc" event={"ID":"d92e5692-ff15-4fc2-98e6-fefdb083f629","Type":"ContainerDied","Data":"eb6cc96a3513a7f1df2a6c1a750904b167d88b502115edfcfcbede9ca76a1afd"} Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.014910 4758 scope.go:117] "RemoveContainer" containerID="d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.014927 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5twc" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.041490 4758 scope.go:117] "RemoveContainer" containerID="687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.082949 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5twc"] Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.088481 4758 scope.go:117] "RemoveContainer" containerID="eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.097252 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r5twc"] Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.114987 4758 scope.go:117] "RemoveContainer" containerID="d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625" Oct 04 11:36:31 crc kubenswrapper[4758]: E1004 11:36:31.115382 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625\": container with ID starting with d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625 not found: ID does not exist" containerID="d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.115418 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625"} err="failed to get container status \"d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625\": rpc error: code = NotFound desc = could not find container \"d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625\": container with ID starting with d8bab2b782525a9c20ea5dd43de099f0481816cc3dc49d8bbcb51a647f726625 not found: ID does not exist" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.115443 4758 scope.go:117] "RemoveContainer" containerID="687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4" Oct 04 11:36:31 crc kubenswrapper[4758]: E1004 11:36:31.115678 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4\": container with ID starting with 687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4 not found: ID does not exist" containerID="687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.115703 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4"} err="failed to get container status \"687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4\": rpc error: code = NotFound desc = could not find container \"687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4\": container with ID starting with 687b01c585da1ca69c4bb97a648a861a53b28cd688e38b9a81ff37b8562c6bb4 not found: ID does not exist" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.115719 4758 scope.go:117] "RemoveContainer" containerID="eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60" Oct 04 11:36:31 crc kubenswrapper[4758]: E1004 11:36:31.115953 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60\": container with ID starting with eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60 not found: ID does not exist" containerID="eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.115979 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60"} err="failed to get container status \"eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60\": rpc error: code = NotFound desc = could not find container \"eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60\": container with ID starting with eb4dd7c0b51b0c08ad622b6dc2504c8a388af5b72862709b568401c0a324cc60 not found: ID does not exist" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.249572 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.249662 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.249761 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.250715 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"86475db01c8dfb4eef260997889015747080409f7c657306fa9c606433685e87"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.250951 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://86475db01c8dfb4eef260997889015747080409f7c657306fa9c606433685e87" gracePeriod=600 Oct 04 11:36:31 crc kubenswrapper[4758]: I1004 11:36:31.339143 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" path="/var/lib/kubelet/pods/d92e5692-ff15-4fc2-98e6-fefdb083f629/volumes" Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.023891 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="86475db01c8dfb4eef260997889015747080409f7c657306fa9c606433685e87" exitCode=0 Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.023948 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"86475db01c8dfb4eef260997889015747080409f7c657306fa9c606433685e87"} Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.024803 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55"} Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.024827 4758 scope.go:117] "RemoveContainer" containerID="636c302f97b74babde7da467e056171a70478464533d342358e91b26e84de888" Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.202874 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.202932 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:32 crc kubenswrapper[4758]: I1004 11:36:32.264513 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:33 crc kubenswrapper[4758]: I1004 11:36:33.099400 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:34 crc kubenswrapper[4758]: I1004 11:36:34.273796 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xcf4"] Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.056660 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6xcf4" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="registry-server" containerID="cri-o://3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78" gracePeriod=2 Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.575886 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.731907 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srdzj\" (UniqueName: \"kubernetes.io/projected/9a6b6bdc-dca8-42b8-bf75-973f402905e9-kube-api-access-srdzj\") pod \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.731972 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-utilities\") pod \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.732149 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-catalog-content\") pod \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\" (UID: \"9a6b6bdc-dca8-42b8-bf75-973f402905e9\") " Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.733234 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-utilities" (OuterVolumeSpecName: "utilities") pod "9a6b6bdc-dca8-42b8-bf75-973f402905e9" (UID: "9a6b6bdc-dca8-42b8-bf75-973f402905e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.738527 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6b6bdc-dca8-42b8-bf75-973f402905e9-kube-api-access-srdzj" (OuterVolumeSpecName: "kube-api-access-srdzj") pod "9a6b6bdc-dca8-42b8-bf75-973f402905e9" (UID: "9a6b6bdc-dca8-42b8-bf75-973f402905e9"). InnerVolumeSpecName "kube-api-access-srdzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.786051 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a6b6bdc-dca8-42b8-bf75-973f402905e9" (UID: "9a6b6bdc-dca8-42b8-bf75-973f402905e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.833384 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srdzj\" (UniqueName: \"kubernetes.io/projected/9a6b6bdc-dca8-42b8-bf75-973f402905e9-kube-api-access-srdzj\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.833417 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:35 crc kubenswrapper[4758]: I1004 11:36:35.833426 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a6b6bdc-dca8-42b8-bf75-973f402905e9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.068137 4758 generic.go:334] "Generic (PLEG): container finished" podID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerID="3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78" exitCode=0 Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.068205 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xcf4" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.068195 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerDied","Data":"3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78"} Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.068771 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xcf4" event={"ID":"9a6b6bdc-dca8-42b8-bf75-973f402905e9","Type":"ContainerDied","Data":"55d314f7c0ad568f97e079b786888efc68a675dd319acdb007ef339d446cb979"} Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.068808 4758 scope.go:117] "RemoveContainer" containerID="3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.099041 4758 scope.go:117] "RemoveContainer" containerID="aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.113223 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xcf4"] Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.131007 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6xcf4"] Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.134347 4758 scope.go:117] "RemoveContainer" containerID="62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.177514 4758 scope.go:117] "RemoveContainer" containerID="3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78" Oct 04 11:36:36 crc kubenswrapper[4758]: E1004 11:36:36.177955 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78\": container with ID starting with 3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78 not found: ID does not exist" containerID="3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.177992 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78"} err="failed to get container status \"3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78\": rpc error: code = NotFound desc = could not find container \"3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78\": container with ID starting with 3253130a5695a400110c3e035b4e82e871ad8b61200e56e463100dac381d3d78 not found: ID does not exist" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.178020 4758 scope.go:117] "RemoveContainer" containerID="aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4" Oct 04 11:36:36 crc kubenswrapper[4758]: E1004 11:36:36.178380 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4\": container with ID starting with aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4 not found: ID does not exist" containerID="aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.178407 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4"} err="failed to get container status \"aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4\": rpc error: code = NotFound desc = could not find container \"aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4\": container with ID starting with aed7085ce7e3c8a1314d9914dc4fd62f13d50e2c4c306bc9f133cd442e1addb4 not found: ID does not exist" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.178425 4758 scope.go:117] "RemoveContainer" containerID="62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54" Oct 04 11:36:36 crc kubenswrapper[4758]: E1004 11:36:36.178889 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54\": container with ID starting with 62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54 not found: ID does not exist" containerID="62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54" Oct 04 11:36:36 crc kubenswrapper[4758]: I1004 11:36:36.178916 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54"} err="failed to get container status \"62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54\": rpc error: code = NotFound desc = could not find container \"62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54\": container with ID starting with 62e57b17bbddd2b4174ced80c57d766523579b18a5b9786455e7b19b2d6d7a54 not found: ID does not exist" Oct 04 11:36:37 crc kubenswrapper[4758]: I1004 11:36:37.344946 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" path="/var/lib/kubelet/pods/9a6b6bdc-dca8-42b8-bf75-973f402905e9/volumes" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.438803 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h9nlk"] Oct 04 11:37:11 crc kubenswrapper[4758]: E1004 11:37:11.440175 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="extract-content" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440200 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="extract-content" Oct 04 11:37:11 crc kubenswrapper[4758]: E1004 11:37:11.440236 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="registry-server" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440249 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="registry-server" Oct 04 11:37:11 crc kubenswrapper[4758]: E1004 11:37:11.440284 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="registry-server" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440298 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="registry-server" Oct 04 11:37:11 crc kubenswrapper[4758]: E1004 11:37:11.440313 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="extract-utilities" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440327 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="extract-utilities" Oct 04 11:37:11 crc kubenswrapper[4758]: E1004 11:37:11.440366 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="extract-utilities" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440378 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="extract-utilities" Oct 04 11:37:11 crc kubenswrapper[4758]: E1004 11:37:11.440408 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="extract-content" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440420 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="extract-content" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440748 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6b6bdc-dca8-42b8-bf75-973f402905e9" containerName="registry-server" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.440805 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d92e5692-ff15-4fc2-98e6-fefdb083f629" containerName="registry-server" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.453411 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.456191 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h9nlk"] Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.553189 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-catalog-content\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.553323 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n92p4\" (UniqueName: \"kubernetes.io/projected/cc66c9f5-d316-4636-8db5-9839391ab1de-kube-api-access-n92p4\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.553349 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-utilities\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.654915 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n92p4\" (UniqueName: \"kubernetes.io/projected/cc66c9f5-d316-4636-8db5-9839391ab1de-kube-api-access-n92p4\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.655162 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-utilities\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.655279 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-catalog-content\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.655723 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-catalog-content\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.656250 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-utilities\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.678181 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n92p4\" (UniqueName: \"kubernetes.io/projected/cc66c9f5-d316-4636-8db5-9839391ab1de-kube-api-access-n92p4\") pod \"community-operators-h9nlk\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:11 crc kubenswrapper[4758]: I1004 11:37:11.799491 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:12 crc kubenswrapper[4758]: I1004 11:37:12.357765 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h9nlk"] Oct 04 11:37:12 crc kubenswrapper[4758]: I1004 11:37:12.509094 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerStarted","Data":"128e2f4bdb6c70710a53dc0853df3cde9868511fdb422b3cbff38da0a88fcdd5"} Oct 04 11:37:13 crc kubenswrapper[4758]: I1004 11:37:13.524771 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerID="ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81" exitCode=0 Oct 04 11:37:13 crc kubenswrapper[4758]: I1004 11:37:13.524826 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerDied","Data":"ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81"} Oct 04 11:37:15 crc kubenswrapper[4758]: I1004 11:37:15.549467 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerStarted","Data":"20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036"} Oct 04 11:37:16 crc kubenswrapper[4758]: I1004 11:37:16.559633 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerID="20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036" exitCode=0 Oct 04 11:37:16 crc kubenswrapper[4758]: I1004 11:37:16.559684 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerDied","Data":"20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036"} Oct 04 11:37:17 crc kubenswrapper[4758]: I1004 11:37:17.569650 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerStarted","Data":"967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa"} Oct 04 11:37:17 crc kubenswrapper[4758]: I1004 11:37:17.601172 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h9nlk" podStartSLOduration=3.158348306 podStartE2EDuration="6.601151779s" podCreationTimestamp="2025-10-04 11:37:11 +0000 UTC" firstStartedPulling="2025-10-04 11:37:13.529021019 +0000 UTC m=+2810.821671948" lastFinishedPulling="2025-10-04 11:37:16.971824532 +0000 UTC m=+2814.264475421" observedRunningTime="2025-10-04 11:37:17.599796703 +0000 UTC m=+2814.892447602" watchObservedRunningTime="2025-10-04 11:37:17.601151779 +0000 UTC m=+2814.893802678" Oct 04 11:37:21 crc kubenswrapper[4758]: I1004 11:37:21.799979 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:21 crc kubenswrapper[4758]: I1004 11:37:21.800760 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:21 crc kubenswrapper[4758]: I1004 11:37:21.852352 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:22 crc kubenswrapper[4758]: I1004 11:37:22.693385 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:22 crc kubenswrapper[4758]: I1004 11:37:22.745855 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h9nlk"] Oct 04 11:37:24 crc kubenswrapper[4758]: I1004 11:37:24.636861 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h9nlk" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="registry-server" containerID="cri-o://967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa" gracePeriod=2 Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.102438 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.257308 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n92p4\" (UniqueName: \"kubernetes.io/projected/cc66c9f5-d316-4636-8db5-9839391ab1de-kube-api-access-n92p4\") pod \"cc66c9f5-d316-4636-8db5-9839391ab1de\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.257831 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-catalog-content\") pod \"cc66c9f5-d316-4636-8db5-9839391ab1de\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.257886 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-utilities\") pod \"cc66c9f5-d316-4636-8db5-9839391ab1de\" (UID: \"cc66c9f5-d316-4636-8db5-9839391ab1de\") " Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.259355 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-utilities" (OuterVolumeSpecName: "utilities") pod "cc66c9f5-d316-4636-8db5-9839391ab1de" (UID: "cc66c9f5-d316-4636-8db5-9839391ab1de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.263326 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc66c9f5-d316-4636-8db5-9839391ab1de-kube-api-access-n92p4" (OuterVolumeSpecName: "kube-api-access-n92p4") pod "cc66c9f5-d316-4636-8db5-9839391ab1de" (UID: "cc66c9f5-d316-4636-8db5-9839391ab1de"). InnerVolumeSpecName "kube-api-access-n92p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.303656 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc66c9f5-d316-4636-8db5-9839391ab1de" (UID: "cc66c9f5-d316-4636-8db5-9839391ab1de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.360839 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.360904 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n92p4\" (UniqueName: \"kubernetes.io/projected/cc66c9f5-d316-4636-8db5-9839391ab1de-kube-api-access-n92p4\") on node \"crc\" DevicePath \"\"" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.360945 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc66c9f5-d316-4636-8db5-9839391ab1de-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.646176 4758 generic.go:334] "Generic (PLEG): container finished" podID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerID="967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa" exitCode=0 Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.646225 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerDied","Data":"967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa"} Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.646260 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9nlk" event={"ID":"cc66c9f5-d316-4636-8db5-9839391ab1de","Type":"ContainerDied","Data":"128e2f4bdb6c70710a53dc0853df3cde9868511fdb422b3cbff38da0a88fcdd5"} Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.646282 4758 scope.go:117] "RemoveContainer" containerID="967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.646429 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9nlk" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.677091 4758 scope.go:117] "RemoveContainer" containerID="20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.677804 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h9nlk"] Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.686224 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h9nlk"] Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.703932 4758 scope.go:117] "RemoveContainer" containerID="ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.754335 4758 scope.go:117] "RemoveContainer" containerID="967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa" Oct 04 11:37:25 crc kubenswrapper[4758]: E1004 11:37:25.754847 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa\": container with ID starting with 967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa not found: ID does not exist" containerID="967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.754885 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa"} err="failed to get container status \"967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa\": rpc error: code = NotFound desc = could not find container \"967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa\": container with ID starting with 967f41dd2ccac7cde976fbe837b7edcb2eb3b8a61cdce80a18b46144f3a924aa not found: ID does not exist" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.754913 4758 scope.go:117] "RemoveContainer" containerID="20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036" Oct 04 11:37:25 crc kubenswrapper[4758]: E1004 11:37:25.755230 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036\": container with ID starting with 20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036 not found: ID does not exist" containerID="20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.755263 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036"} err="failed to get container status \"20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036\": rpc error: code = NotFound desc = could not find container \"20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036\": container with ID starting with 20ebe28b0daa18869cf34d331fd504859be935f971e080c0146a94ca1c775036 not found: ID does not exist" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.755283 4758 scope.go:117] "RemoveContainer" containerID="ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81" Oct 04 11:37:25 crc kubenswrapper[4758]: E1004 11:37:25.755563 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81\": container with ID starting with ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81 not found: ID does not exist" containerID="ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81" Oct 04 11:37:25 crc kubenswrapper[4758]: I1004 11:37:25.755588 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81"} err="failed to get container status \"ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81\": rpc error: code = NotFound desc = could not find container \"ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81\": container with ID starting with ad7914ad02d31932e8f4586a6b6d7b415273a76632ce8c1f15061f39c3f7eb81 not found: ID does not exist" Oct 04 11:37:27 crc kubenswrapper[4758]: I1004 11:37:27.336721 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" path="/var/lib/kubelet/pods/cc66c9f5-d316-4636-8db5-9839391ab1de/volumes" Oct 04 11:38:31 crc kubenswrapper[4758]: I1004 11:38:31.250169 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:38:31 crc kubenswrapper[4758]: I1004 11:38:31.252453 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:38:40 crc kubenswrapper[4758]: I1004 11:38:40.342765 4758 generic.go:334] "Generic (PLEG): container finished" podID="acd41b88-adbc-46b6-b2f0-dde72b71dbf5" containerID="abe8b5ab1106b7089cbfa7c06e70acfff41c77bba6f2d80d594957df58c2457a" exitCode=0 Oct 04 11:38:40 crc kubenswrapper[4758]: I1004 11:38:40.342870 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" event={"ID":"acd41b88-adbc-46b6-b2f0-dde72b71dbf5","Type":"ContainerDied","Data":"abe8b5ab1106b7089cbfa7c06e70acfff41c77bba6f2d80d594957df58c2457a"} Oct 04 11:38:41 crc kubenswrapper[4758]: I1004 11:38:41.839825 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013544 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-4\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-4\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013639 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-telemetry-combined-ca-bundle\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013713 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-0\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013757 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-2\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013823 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w4ml\" (UniqueName: \"kubernetes.io/projected/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-kube-api-access-7w4ml\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013861 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ssh-key\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013895 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-inventory\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013927 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-1\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.013969 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-3\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-3\") pod \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\" (UID: \"acd41b88-adbc-46b6-b2f0-dde72b71dbf5\") " Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.025616 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-kube-api-access-7w4ml" (OuterVolumeSpecName: "kube-api-access-7w4ml") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "kube-api-access-7w4ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.044245 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-inventory" (OuterVolumeSpecName: "inventory") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.045189 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.046182 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.049473 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-3" (OuterVolumeSpecName: "ceilometer-compute-config-data-3") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "ceilometer-compute-config-data-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.061347 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-4" (OuterVolumeSpecName: "ceilometer-compute-config-data-4") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "ceilometer-compute-config-data-4". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.064406 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.072709 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.078838 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "acd41b88-adbc-46b6-b2f0-dde72b71dbf5" (UID: "acd41b88-adbc-46b6-b2f0-dde72b71dbf5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115843 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-3\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-3\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115878 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-4\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-4\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115889 4758 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115899 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115909 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115919 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w4ml\" (UniqueName: \"kubernetes.io/projected/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-kube-api-access-7w4ml\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115929 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115938 4758 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.115947 4758 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/acd41b88-adbc-46b6-b2f0-dde72b71dbf5-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.367172 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" event={"ID":"acd41b88-adbc-46b6-b2f0-dde72b71dbf5","Type":"ContainerDied","Data":"c3839868e774c22a954803fb030f62032149d6893b79b84a322cbb2be18adc22"} Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.367226 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3839868e774c22a954803fb030f62032149d6893b79b84a322cbb2be18adc22" Oct 04 11:38:42 crc kubenswrapper[4758]: I1004 11:38:42.367225 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs" Oct 04 11:39:01 crc kubenswrapper[4758]: I1004 11:39:01.249975 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:39:01 crc kubenswrapper[4758]: I1004 11:39:01.250560 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.454340 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfrh"] Oct 04 11:39:26 crc kubenswrapper[4758]: E1004 11:39:26.455244 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="extract-utilities" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.455257 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="extract-utilities" Oct 04 11:39:26 crc kubenswrapper[4758]: E1004 11:39:26.455271 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd41b88-adbc-46b6-b2f0-dde72b71dbf5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.455280 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd41b88-adbc-46b6-b2f0-dde72b71dbf5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 11:39:26 crc kubenswrapper[4758]: E1004 11:39:26.455296 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="extract-content" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.455304 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="extract-content" Oct 04 11:39:26 crc kubenswrapper[4758]: E1004 11:39:26.455325 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="registry-server" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.455331 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="registry-server" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.455518 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc66c9f5-d316-4636-8db5-9839391ab1de" containerName="registry-server" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.455567 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd41b88-adbc-46b6-b2f0-dde72b71dbf5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.456956 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.474254 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfrh"] Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.617229 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljrqd\" (UniqueName: \"kubernetes.io/projected/02e894d4-f647-4d7e-a797-b22353dff049-kube-api-access-ljrqd\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.617332 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-utilities\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.617370 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-catalog-content\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.718641 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljrqd\" (UniqueName: \"kubernetes.io/projected/02e894d4-f647-4d7e-a797-b22353dff049-kube-api-access-ljrqd\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.718775 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-utilities\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.718822 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-catalog-content\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.719253 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-utilities\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.719341 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-catalog-content\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.742123 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljrqd\" (UniqueName: \"kubernetes.io/projected/02e894d4-f647-4d7e-a797-b22353dff049-kube-api-access-ljrqd\") pod \"redhat-marketplace-2qfrh\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:26 crc kubenswrapper[4758]: I1004 11:39:26.783639 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:27 crc kubenswrapper[4758]: I1004 11:39:27.263473 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfrh"] Oct 04 11:39:27 crc kubenswrapper[4758]: I1004 11:39:27.843019 4758 generic.go:334] "Generic (PLEG): container finished" podID="02e894d4-f647-4d7e-a797-b22353dff049" containerID="1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9" exitCode=0 Oct 04 11:39:27 crc kubenswrapper[4758]: I1004 11:39:27.843097 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfrh" event={"ID":"02e894d4-f647-4d7e-a797-b22353dff049","Type":"ContainerDied","Data":"1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9"} Oct 04 11:39:27 crc kubenswrapper[4758]: I1004 11:39:27.844282 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfrh" event={"ID":"02e894d4-f647-4d7e-a797-b22353dff049","Type":"ContainerStarted","Data":"f060f800def7bbba46a757de3d05461da56a5e1e0c6ca7aaebce79464f5fe1ed"} Oct 04 11:39:29 crc kubenswrapper[4758]: I1004 11:39:29.858695 4758 generic.go:334] "Generic (PLEG): container finished" podID="02e894d4-f647-4d7e-a797-b22353dff049" containerID="3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf" exitCode=0 Oct 04 11:39:29 crc kubenswrapper[4758]: I1004 11:39:29.858931 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfrh" event={"ID":"02e894d4-f647-4d7e-a797-b22353dff049","Type":"ContainerDied","Data":"3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf"} Oct 04 11:39:30 crc kubenswrapper[4758]: I1004 11:39:30.869982 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfrh" event={"ID":"02e894d4-f647-4d7e-a797-b22353dff049","Type":"ContainerStarted","Data":"5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2"} Oct 04 11:39:30 crc kubenswrapper[4758]: I1004 11:39:30.900442 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2qfrh" podStartSLOduration=2.410916651 podStartE2EDuration="4.900421991s" podCreationTimestamp="2025-10-04 11:39:26 +0000 UTC" firstStartedPulling="2025-10-04 11:39:27.846034644 +0000 UTC m=+2945.138685533" lastFinishedPulling="2025-10-04 11:39:30.335539984 +0000 UTC m=+2947.628190873" observedRunningTime="2025-10-04 11:39:30.889561861 +0000 UTC m=+2948.182212770" watchObservedRunningTime="2025-10-04 11:39:30.900421991 +0000 UTC m=+2948.193072890" Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.249557 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.249625 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.249669 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.250327 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.250380 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" gracePeriod=600 Oct 04 11:39:31 crc kubenswrapper[4758]: E1004 11:39:31.400967 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.881359 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" exitCode=0 Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.882145 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55"} Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.882181 4758 scope.go:117] "RemoveContainer" containerID="86475db01c8dfb4eef260997889015747080409f7c657306fa9c606433685e87" Oct 04 11:39:31 crc kubenswrapper[4758]: I1004 11:39:31.882468 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:39:31 crc kubenswrapper[4758]: E1004 11:39:31.882663 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:39:36 crc kubenswrapper[4758]: I1004 11:39:36.784343 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:36 crc kubenswrapper[4758]: I1004 11:39:36.784981 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:36 crc kubenswrapper[4758]: I1004 11:39:36.848760 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:37 crc kubenswrapper[4758]: I1004 11:39:37.017647 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:37 crc kubenswrapper[4758]: I1004 11:39:37.088944 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfrh"] Oct 04 11:39:38 crc kubenswrapper[4758]: I1004 11:39:38.996873 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2qfrh" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="registry-server" containerID="cri-o://5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2" gracePeriod=2 Oct 04 11:39:39 crc kubenswrapper[4758]: I1004 11:39:39.920396 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.007800 4758 generic.go:334] "Generic (PLEG): container finished" podID="02e894d4-f647-4d7e-a797-b22353dff049" containerID="5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2" exitCode=0 Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.007868 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qfrh" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.007889 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfrh" event={"ID":"02e894d4-f647-4d7e-a797-b22353dff049","Type":"ContainerDied","Data":"5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2"} Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.008617 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qfrh" event={"ID":"02e894d4-f647-4d7e-a797-b22353dff049","Type":"ContainerDied","Data":"f060f800def7bbba46a757de3d05461da56a5e1e0c6ca7aaebce79464f5fe1ed"} Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.008643 4758 scope.go:117] "RemoveContainer" containerID="5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.032184 4758 scope.go:117] "RemoveContainer" containerID="3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.054158 4758 scope.go:117] "RemoveContainer" containerID="1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.070261 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-catalog-content\") pod \"02e894d4-f647-4d7e-a797-b22353dff049\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.070350 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-utilities\") pod \"02e894d4-f647-4d7e-a797-b22353dff049\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.070403 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljrqd\" (UniqueName: \"kubernetes.io/projected/02e894d4-f647-4d7e-a797-b22353dff049-kube-api-access-ljrqd\") pod \"02e894d4-f647-4d7e-a797-b22353dff049\" (UID: \"02e894d4-f647-4d7e-a797-b22353dff049\") " Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.071538 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-utilities" (OuterVolumeSpecName: "utilities") pod "02e894d4-f647-4d7e-a797-b22353dff049" (UID: "02e894d4-f647-4d7e-a797-b22353dff049"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.077335 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e894d4-f647-4d7e-a797-b22353dff049-kube-api-access-ljrqd" (OuterVolumeSpecName: "kube-api-access-ljrqd") pod "02e894d4-f647-4d7e-a797-b22353dff049" (UID: "02e894d4-f647-4d7e-a797-b22353dff049"). InnerVolumeSpecName "kube-api-access-ljrqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.086646 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02e894d4-f647-4d7e-a797-b22353dff049" (UID: "02e894d4-f647-4d7e-a797-b22353dff049"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.137358 4758 scope.go:117] "RemoveContainer" containerID="5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2" Oct 04 11:39:40 crc kubenswrapper[4758]: E1004 11:39:40.137813 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2\": container with ID starting with 5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2 not found: ID does not exist" containerID="5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.137861 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2"} err="failed to get container status \"5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2\": rpc error: code = NotFound desc = could not find container \"5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2\": container with ID starting with 5b37e5d0b5e0dadebc35a571b3ec5f14487799ad83dc6535c9acf5ed4014f9f2 not found: ID does not exist" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.137893 4758 scope.go:117] "RemoveContainer" containerID="3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf" Oct 04 11:39:40 crc kubenswrapper[4758]: E1004 11:39:40.138516 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf\": container with ID starting with 3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf not found: ID does not exist" containerID="3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.138545 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf"} err="failed to get container status \"3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf\": rpc error: code = NotFound desc = could not find container \"3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf\": container with ID starting with 3986d255d1f473908753cd57e4e8c46b76260d69930fc8e032d092470b4c8abf not found: ID does not exist" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.138563 4758 scope.go:117] "RemoveContainer" containerID="1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9" Oct 04 11:39:40 crc kubenswrapper[4758]: E1004 11:39:40.139081 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9\": container with ID starting with 1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9 not found: ID does not exist" containerID="1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.139124 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9"} err="failed to get container status \"1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9\": rpc error: code = NotFound desc = could not find container \"1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9\": container with ID starting with 1ca5593e91edffaf7f7504a148a4469e67f1f8c14a3eaa1ce91edbb541ca9ae9 not found: ID does not exist" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.172809 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.173232 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02e894d4-f647-4d7e-a797-b22353dff049-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.173250 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljrqd\" (UniqueName: \"kubernetes.io/projected/02e894d4-f647-4d7e-a797-b22353dff049-kube-api-access-ljrqd\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.345950 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfrh"] Oct 04 11:39:40 crc kubenswrapper[4758]: I1004 11:39:40.357698 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qfrh"] Oct 04 11:39:41 crc kubenswrapper[4758]: I1004 11:39:41.341280 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02e894d4-f647-4d7e-a797-b22353dff049" path="/var/lib/kubelet/pods/02e894d4-f647-4d7e-a797-b22353dff049/volumes" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.407940 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 11:39:43 crc kubenswrapper[4758]: E1004 11:39:43.408648 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="extract-content" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.408663 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="extract-content" Oct 04 11:39:43 crc kubenswrapper[4758]: E1004 11:39:43.408688 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="registry-server" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.408694 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="registry-server" Oct 04 11:39:43 crc kubenswrapper[4758]: E1004 11:39:43.408725 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="extract-utilities" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.408733 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="extract-utilities" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.408891 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="02e894d4-f647-4d7e-a797-b22353dff049" containerName="registry-server" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.409527 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.411478 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.411771 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.415861 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.415979 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9jwgs" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.423147 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535474 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535651 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535694 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-config-data\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535764 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535794 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535829 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535883 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.535920 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.536145 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ddl8\" (UniqueName: \"kubernetes.io/projected/8a041496-5b8c-4d15-961a-d710b5a461fe-kube-api-access-5ddl8\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638020 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ddl8\" (UniqueName: \"kubernetes.io/projected/8a041496-5b8c-4d15-961a-d710b5a461fe-kube-api-access-5ddl8\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638333 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638432 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638510 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-config-data\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638595 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638660 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638735 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638813 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638887 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.638906 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.639375 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.639995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.641222 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-config-data\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.642995 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.646225 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.646221 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.647970 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.668421 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ddl8\" (UniqueName: \"kubernetes.io/projected/8a041496-5b8c-4d15-961a-d710b5a461fe-kube-api-access-5ddl8\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.674181 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " pod="openstack/tempest-tests-tempest" Oct 04 11:39:43 crc kubenswrapper[4758]: I1004 11:39:43.732972 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:39:44 crc kubenswrapper[4758]: I1004 11:39:44.206771 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 11:39:44 crc kubenswrapper[4758]: W1004 11:39:44.218020 4758 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a041496_5b8c_4d15_961a_d710b5a461fe.slice/crio-2506b20b0aab557c3a1a225dade4a5bfe186e9deba016fe5d809372c1632d557 WatchSource:0}: Error finding container 2506b20b0aab557c3a1a225dade4a5bfe186e9deba016fe5d809372c1632d557: Status 404 returned error can't find the container with id 2506b20b0aab557c3a1a225dade4a5bfe186e9deba016fe5d809372c1632d557 Oct 04 11:39:45 crc kubenswrapper[4758]: I1004 11:39:45.056830 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8a041496-5b8c-4d15-961a-d710b5a461fe","Type":"ContainerStarted","Data":"2506b20b0aab557c3a1a225dade4a5bfe186e9deba016fe5d809372c1632d557"} Oct 04 11:39:46 crc kubenswrapper[4758]: I1004 11:39:46.328333 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:39:46 crc kubenswrapper[4758]: E1004 11:39:46.328649 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:39:57 crc kubenswrapper[4758]: I1004 11:39:57.326661 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:39:57 crc kubenswrapper[4758]: E1004 11:39:57.327446 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:40:10 crc kubenswrapper[4758]: I1004 11:40:10.332206 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:40:10 crc kubenswrapper[4758]: E1004 11:40:10.333362 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:40:15 crc kubenswrapper[4758]: E1004 11:40:15.441134 4758 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 04 11:40:15 crc kubenswrapper[4758]: E1004 11:40:15.444381 4758 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5ddl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(8a041496-5b8c-4d15-961a-d710b5a461fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:40:15 crc kubenswrapper[4758]: E1004 11:40:15.445544 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="8a041496-5b8c-4d15-961a-d710b5a461fe" Oct 04 11:40:16 crc kubenswrapper[4758]: E1004 11:40:16.431012 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="8a041496-5b8c-4d15-961a-d710b5a461fe" Oct 04 11:40:25 crc kubenswrapper[4758]: I1004 11:40:25.326233 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:40:25 crc kubenswrapper[4758]: E1004 11:40:25.327373 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:40:30 crc kubenswrapper[4758]: I1004 11:40:30.329546 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:40:30 crc kubenswrapper[4758]: I1004 11:40:30.779937 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 11:40:32 crc kubenswrapper[4758]: I1004 11:40:32.594968 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8a041496-5b8c-4d15-961a-d710b5a461fe","Type":"ContainerStarted","Data":"db31500cd99d82268b9b85556f8fd21ca59e05dcfca81fd0d345c081052a530f"} Oct 04 11:40:32 crc kubenswrapper[4758]: I1004 11:40:32.615001 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.058593918 podStartE2EDuration="50.614965664s" podCreationTimestamp="2025-10-04 11:39:42 +0000 UTC" firstStartedPulling="2025-10-04 11:39:44.220639097 +0000 UTC m=+2961.513289986" lastFinishedPulling="2025-10-04 11:40:30.777010833 +0000 UTC m=+3008.069661732" observedRunningTime="2025-10-04 11:40:32.613162126 +0000 UTC m=+3009.905813015" watchObservedRunningTime="2025-10-04 11:40:32.614965664 +0000 UTC m=+3009.907616573" Oct 04 11:40:40 crc kubenswrapper[4758]: I1004 11:40:40.326778 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:40:40 crc kubenswrapper[4758]: E1004 11:40:40.327701 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:40:55 crc kubenswrapper[4758]: I1004 11:40:55.326351 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:40:55 crc kubenswrapper[4758]: E1004 11:40:55.326978 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:41:07 crc kubenswrapper[4758]: I1004 11:41:07.329971 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:41:07 crc kubenswrapper[4758]: E1004 11:41:07.330772 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:41:22 crc kubenswrapper[4758]: I1004 11:41:22.326952 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:41:22 crc kubenswrapper[4758]: E1004 11:41:22.327678 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:41:37 crc kubenswrapper[4758]: I1004 11:41:37.329730 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:41:37 crc kubenswrapper[4758]: E1004 11:41:37.330386 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:41:49 crc kubenswrapper[4758]: I1004 11:41:49.329220 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:41:49 crc kubenswrapper[4758]: E1004 11:41:49.329921 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:42:02 crc kubenswrapper[4758]: I1004 11:42:02.326402 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:42:02 crc kubenswrapper[4758]: E1004 11:42:02.327375 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:42:15 crc kubenswrapper[4758]: I1004 11:42:15.329897 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:42:15 crc kubenswrapper[4758]: E1004 11:42:15.331879 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:42:30 crc kubenswrapper[4758]: I1004 11:42:30.325950 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:42:30 crc kubenswrapper[4758]: E1004 11:42:30.327026 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:42:45 crc kubenswrapper[4758]: I1004 11:42:45.325726 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:42:45 crc kubenswrapper[4758]: E1004 11:42:45.326457 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:42:58 crc kubenswrapper[4758]: I1004 11:42:58.326307 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:42:58 crc kubenswrapper[4758]: E1004 11:42:58.326982 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:43:12 crc kubenswrapper[4758]: I1004 11:43:12.326390 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:43:12 crc kubenswrapper[4758]: E1004 11:43:12.327133 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:43:26 crc kubenswrapper[4758]: I1004 11:43:26.325961 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:43:26 crc kubenswrapper[4758]: E1004 11:43:26.326714 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:43:40 crc kubenswrapper[4758]: I1004 11:43:40.327530 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:43:40 crc kubenswrapper[4758]: E1004 11:43:40.328383 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:43:51 crc kubenswrapper[4758]: I1004 11:43:51.325626 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:43:51 crc kubenswrapper[4758]: E1004 11:43:51.326248 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:44:05 crc kubenswrapper[4758]: I1004 11:44:05.326671 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:44:05 crc kubenswrapper[4758]: E1004 11:44:05.328652 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:44:19 crc kubenswrapper[4758]: I1004 11:44:19.325946 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:44:19 crc kubenswrapper[4758]: E1004 11:44:19.327003 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:44:33 crc kubenswrapper[4758]: I1004 11:44:33.326521 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:44:33 crc kubenswrapper[4758]: I1004 11:44:33.708601 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"bdbd2a04ee42e03967d4eca5f189346fe862da26e4308abe36307465a4dcbbcc"} Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.178415 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6"] Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.180369 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.183245 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.184377 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.209675 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6"] Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.300954 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0fe94547-703e-4398-81da-f97ebfaf2ff9-config-volume\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.301233 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqmm7\" (UniqueName: \"kubernetes.io/projected/0fe94547-703e-4398-81da-f97ebfaf2ff9-kube-api-access-lqmm7\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.301364 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0fe94547-703e-4398-81da-f97ebfaf2ff9-secret-volume\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.403662 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0fe94547-703e-4398-81da-f97ebfaf2ff9-config-volume\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.404214 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqmm7\" (UniqueName: \"kubernetes.io/projected/0fe94547-703e-4398-81da-f97ebfaf2ff9-kube-api-access-lqmm7\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.404349 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0fe94547-703e-4398-81da-f97ebfaf2ff9-secret-volume\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.404764 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0fe94547-703e-4398-81da-f97ebfaf2ff9-config-volume\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.411576 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0fe94547-703e-4398-81da-f97ebfaf2ff9-secret-volume\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.422491 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqmm7\" (UniqueName: \"kubernetes.io/projected/0fe94547-703e-4398-81da-f97ebfaf2ff9-kube-api-access-lqmm7\") pod \"collect-profiles-29326305-pz6m6\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:00 crc kubenswrapper[4758]: I1004 11:45:00.500008 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:01 crc kubenswrapper[4758]: I1004 11:45:01.533203 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6"] Oct 04 11:45:01 crc kubenswrapper[4758]: I1004 11:45:01.966771 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" event={"ID":"0fe94547-703e-4398-81da-f97ebfaf2ff9","Type":"ContainerStarted","Data":"c2a0dae9d37428c2c34ba4ed37b5642824ed71f562ab5e2a69dc15387af99f35"} Oct 04 11:45:01 crc kubenswrapper[4758]: I1004 11:45:01.966815 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" event={"ID":"0fe94547-703e-4398-81da-f97ebfaf2ff9","Type":"ContainerStarted","Data":"329b71f53adc7e44d8015a03a162018b0665efa284942ee6501d4ff23789d64b"} Oct 04 11:45:01 crc kubenswrapper[4758]: I1004 11:45:01.985510 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" podStartSLOduration=1.985491672 podStartE2EDuration="1.985491672s" podCreationTimestamp="2025-10-04 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:45:01.979983145 +0000 UTC m=+3279.272634044" watchObservedRunningTime="2025-10-04 11:45:01.985491672 +0000 UTC m=+3279.278142561" Oct 04 11:45:02 crc kubenswrapper[4758]: I1004 11:45:02.978761 4758 generic.go:334] "Generic (PLEG): container finished" podID="0fe94547-703e-4398-81da-f97ebfaf2ff9" containerID="c2a0dae9d37428c2c34ba4ed37b5642824ed71f562ab5e2a69dc15387af99f35" exitCode=0 Oct 04 11:45:02 crc kubenswrapper[4758]: I1004 11:45:02.978947 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" event={"ID":"0fe94547-703e-4398-81da-f97ebfaf2ff9","Type":"ContainerDied","Data":"c2a0dae9d37428c2c34ba4ed37b5642824ed71f562ab5e2a69dc15387af99f35"} Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.362965 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.488590 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqmm7\" (UniqueName: \"kubernetes.io/projected/0fe94547-703e-4398-81da-f97ebfaf2ff9-kube-api-access-lqmm7\") pod \"0fe94547-703e-4398-81da-f97ebfaf2ff9\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.488630 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0fe94547-703e-4398-81da-f97ebfaf2ff9-secret-volume\") pod \"0fe94547-703e-4398-81da-f97ebfaf2ff9\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.488675 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0fe94547-703e-4398-81da-f97ebfaf2ff9-config-volume\") pod \"0fe94547-703e-4398-81da-f97ebfaf2ff9\" (UID: \"0fe94547-703e-4398-81da-f97ebfaf2ff9\") " Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.489757 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fe94547-703e-4398-81da-f97ebfaf2ff9-config-volume" (OuterVolumeSpecName: "config-volume") pod "0fe94547-703e-4398-81da-f97ebfaf2ff9" (UID: "0fe94547-703e-4398-81da-f97ebfaf2ff9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.495259 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fe94547-703e-4398-81da-f97ebfaf2ff9-kube-api-access-lqmm7" (OuterVolumeSpecName: "kube-api-access-lqmm7") pod "0fe94547-703e-4398-81da-f97ebfaf2ff9" (UID: "0fe94547-703e-4398-81da-f97ebfaf2ff9"). InnerVolumeSpecName "kube-api-access-lqmm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.503252 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fe94547-703e-4398-81da-f97ebfaf2ff9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0fe94547-703e-4398-81da-f97ebfaf2ff9" (UID: "0fe94547-703e-4398-81da-f97ebfaf2ff9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.591617 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqmm7\" (UniqueName: \"kubernetes.io/projected/0fe94547-703e-4398-81da-f97ebfaf2ff9-kube-api-access-lqmm7\") on node \"crc\" DevicePath \"\"" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.592253 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0fe94547-703e-4398-81da-f97ebfaf2ff9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.592388 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0fe94547-703e-4398-81da-f97ebfaf2ff9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.610907 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2"] Oct 04 11:45:04 crc kubenswrapper[4758]: I1004 11:45:04.616945 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-n59c2"] Oct 04 11:45:05 crc kubenswrapper[4758]: I1004 11:45:05.000031 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" event={"ID":"0fe94547-703e-4398-81da-f97ebfaf2ff9","Type":"ContainerDied","Data":"329b71f53adc7e44d8015a03a162018b0665efa284942ee6501d4ff23789d64b"} Oct 04 11:45:05 crc kubenswrapper[4758]: I1004 11:45:05.000093 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="329b71f53adc7e44d8015a03a162018b0665efa284942ee6501d4ff23789d64b" Oct 04 11:45:05 crc kubenswrapper[4758]: I1004 11:45:05.000091 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-pz6m6" Oct 04 11:45:05 crc kubenswrapper[4758]: I1004 11:45:05.342786 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c2c19f-863a-4faf-9818-1ce9ab1ada00" path="/var/lib/kubelet/pods/54c2c19f-863a-4faf-9818-1ce9ab1ada00/volumes" Oct 04 11:45:15 crc kubenswrapper[4758]: I1004 11:45:15.471055 4758 scope.go:117] "RemoveContainer" containerID="a7a604075251b79e81ab1b814cb2867b52991cd45ca8cb8d33bf64f89a3440cb" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.461439 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pthxf"] Oct 04 11:46:22 crc kubenswrapper[4758]: E1004 11:46:22.462203 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fe94547-703e-4398-81da-f97ebfaf2ff9" containerName="collect-profiles" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.462216 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fe94547-703e-4398-81da-f97ebfaf2ff9" containerName="collect-profiles" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.462405 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fe94547-703e-4398-81da-f97ebfaf2ff9" containerName="collect-profiles" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.463718 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.481885 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pthxf"] Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.584807 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-utilities\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.585023 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj99d\" (UniqueName: \"kubernetes.io/projected/2324546b-1772-42a6-b0ec-87811aebc102-kube-api-access-fj99d\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.585049 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-catalog-content\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.686657 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-utilities\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.686803 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-catalog-content\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.686821 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj99d\" (UniqueName: \"kubernetes.io/projected/2324546b-1772-42a6-b0ec-87811aebc102-kube-api-access-fj99d\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.687272 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-utilities\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.687420 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-catalog-content\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.725113 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj99d\" (UniqueName: \"kubernetes.io/projected/2324546b-1772-42a6-b0ec-87811aebc102-kube-api-access-fj99d\") pod \"redhat-operators-pthxf\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:22 crc kubenswrapper[4758]: I1004 11:46:22.783584 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:23 crc kubenswrapper[4758]: I1004 11:46:23.295556 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pthxf"] Oct 04 11:46:23 crc kubenswrapper[4758]: I1004 11:46:23.666630 4758 generic.go:334] "Generic (PLEG): container finished" podID="2324546b-1772-42a6-b0ec-87811aebc102" containerID="ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26" exitCode=0 Oct 04 11:46:23 crc kubenswrapper[4758]: I1004 11:46:23.666849 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerDied","Data":"ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26"} Oct 04 11:46:23 crc kubenswrapper[4758]: I1004 11:46:23.666923 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerStarted","Data":"e6da4ecc89d96f20856c8473d29dd65bae078cb606b2cd2c75bddb35290be60f"} Oct 04 11:46:23 crc kubenswrapper[4758]: I1004 11:46:23.668839 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:46:24 crc kubenswrapper[4758]: I1004 11:46:24.676442 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerStarted","Data":"6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d"} Oct 04 11:46:29 crc kubenswrapper[4758]: I1004 11:46:29.722808 4758 generic.go:334] "Generic (PLEG): container finished" podID="2324546b-1772-42a6-b0ec-87811aebc102" containerID="6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d" exitCode=0 Oct 04 11:46:29 crc kubenswrapper[4758]: I1004 11:46:29.722901 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerDied","Data":"6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d"} Oct 04 11:46:30 crc kubenswrapper[4758]: I1004 11:46:30.733466 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerStarted","Data":"435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c"} Oct 04 11:46:30 crc kubenswrapper[4758]: I1004 11:46:30.754896 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pthxf" podStartSLOduration=1.974540728 podStartE2EDuration="8.754874753s" podCreationTimestamp="2025-10-04 11:46:22 +0000 UTC" firstStartedPulling="2025-10-04 11:46:23.668571597 +0000 UTC m=+3360.961222486" lastFinishedPulling="2025-10-04 11:46:30.448905612 +0000 UTC m=+3367.741556511" observedRunningTime="2025-10-04 11:46:30.74725855 +0000 UTC m=+3368.039909439" watchObservedRunningTime="2025-10-04 11:46:30.754874753 +0000 UTC m=+3368.047525642" Oct 04 11:46:32 crc kubenswrapper[4758]: I1004 11:46:32.784358 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:32 crc kubenswrapper[4758]: I1004 11:46:32.784745 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:33 crc kubenswrapper[4758]: I1004 11:46:33.845939 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pthxf" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="registry-server" probeResult="failure" output=< Oct 04 11:46:33 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:46:33 crc kubenswrapper[4758]: > Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.371080 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dnwbs"] Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.373209 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.382636 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnwbs"] Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.501023 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-utilities\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.501064 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsbd4\" (UniqueName: \"kubernetes.io/projected/f8ebb362-6343-414e-9369-bb24fa7411a0-kube-api-access-bsbd4\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.501197 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-catalog-content\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.602607 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-utilities\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.602921 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsbd4\" (UniqueName: \"kubernetes.io/projected/f8ebb362-6343-414e-9369-bb24fa7411a0-kube-api-access-bsbd4\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.602988 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-catalog-content\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.603080 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-utilities\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.603359 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-catalog-content\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.624113 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsbd4\" (UniqueName: \"kubernetes.io/projected/f8ebb362-6343-414e-9369-bb24fa7411a0-kube-api-access-bsbd4\") pod \"certified-operators-dnwbs\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:34 crc kubenswrapper[4758]: I1004 11:46:34.700344 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:35 crc kubenswrapper[4758]: I1004 11:46:35.252388 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnwbs"] Oct 04 11:46:35 crc kubenswrapper[4758]: I1004 11:46:35.776633 4758 generic.go:334] "Generic (PLEG): container finished" podID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerID="25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e" exitCode=0 Oct 04 11:46:35 crc kubenswrapper[4758]: I1004 11:46:35.776678 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerDied","Data":"25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e"} Oct 04 11:46:35 crc kubenswrapper[4758]: I1004 11:46:35.776704 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerStarted","Data":"addb425b2b16215b53f73766f004d443d5947345d06de19431a19dcb33bce4c9"} Oct 04 11:46:36 crc kubenswrapper[4758]: I1004 11:46:36.791698 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerStarted","Data":"e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d"} Oct 04 11:46:38 crc kubenswrapper[4758]: I1004 11:46:38.807092 4758 generic.go:334] "Generic (PLEG): container finished" podID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerID="e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d" exitCode=0 Oct 04 11:46:38 crc kubenswrapper[4758]: I1004 11:46:38.807142 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerDied","Data":"e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d"} Oct 04 11:46:39 crc kubenswrapper[4758]: I1004 11:46:39.817895 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerStarted","Data":"3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba"} Oct 04 11:46:39 crc kubenswrapper[4758]: I1004 11:46:39.838228 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dnwbs" podStartSLOduration=2.383308355 podStartE2EDuration="5.838209254s" podCreationTimestamp="2025-10-04 11:46:34 +0000 UTC" firstStartedPulling="2025-10-04 11:46:35.778270517 +0000 UTC m=+3373.070921396" lastFinishedPulling="2025-10-04 11:46:39.233171406 +0000 UTC m=+3376.525822295" observedRunningTime="2025-10-04 11:46:39.834291139 +0000 UTC m=+3377.126942028" watchObservedRunningTime="2025-10-04 11:46:39.838209254 +0000 UTC m=+3377.130860143" Oct 04 11:46:43 crc kubenswrapper[4758]: I1004 11:46:43.852049 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pthxf" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="registry-server" probeResult="failure" output=< Oct 04 11:46:43 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:46:43 crc kubenswrapper[4758]: > Oct 04 11:46:44 crc kubenswrapper[4758]: I1004 11:46:44.701138 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:44 crc kubenswrapper[4758]: I1004 11:46:44.701180 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:44 crc kubenswrapper[4758]: I1004 11:46:44.765810 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:44 crc kubenswrapper[4758]: I1004 11:46:44.897896 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:45 crc kubenswrapper[4758]: I1004 11:46:45.004753 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnwbs"] Oct 04 11:46:46 crc kubenswrapper[4758]: I1004 11:46:46.873444 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dnwbs" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="registry-server" containerID="cri-o://3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba" gracePeriod=2 Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.400394 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.469959 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsbd4\" (UniqueName: \"kubernetes.io/projected/f8ebb362-6343-414e-9369-bb24fa7411a0-kube-api-access-bsbd4\") pod \"f8ebb362-6343-414e-9369-bb24fa7411a0\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.470388 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-utilities\") pod \"f8ebb362-6343-414e-9369-bb24fa7411a0\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.470694 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-catalog-content\") pod \"f8ebb362-6343-414e-9369-bb24fa7411a0\" (UID: \"f8ebb362-6343-414e-9369-bb24fa7411a0\") " Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.470944 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-utilities" (OuterVolumeSpecName: "utilities") pod "f8ebb362-6343-414e-9369-bb24fa7411a0" (UID: "f8ebb362-6343-414e-9369-bb24fa7411a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.471381 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.481292 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8ebb362-6343-414e-9369-bb24fa7411a0-kube-api-access-bsbd4" (OuterVolumeSpecName: "kube-api-access-bsbd4") pod "f8ebb362-6343-414e-9369-bb24fa7411a0" (UID: "f8ebb362-6343-414e-9369-bb24fa7411a0"). InnerVolumeSpecName "kube-api-access-bsbd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.512075 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8ebb362-6343-414e-9369-bb24fa7411a0" (UID: "f8ebb362-6343-414e-9369-bb24fa7411a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.573244 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsbd4\" (UniqueName: \"kubernetes.io/projected/f8ebb362-6343-414e-9369-bb24fa7411a0-kube-api-access-bsbd4\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.573305 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ebb362-6343-414e-9369-bb24fa7411a0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.882778 4758 generic.go:334] "Generic (PLEG): container finished" podID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerID="3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba" exitCode=0 Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.882822 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerDied","Data":"3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba"} Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.882849 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnwbs" event={"ID":"f8ebb362-6343-414e-9369-bb24fa7411a0","Type":"ContainerDied","Data":"addb425b2b16215b53f73766f004d443d5947345d06de19431a19dcb33bce4c9"} Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.882866 4758 scope.go:117] "RemoveContainer" containerID="3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.882899 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnwbs" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.914799 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnwbs"] Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.918894 4758 scope.go:117] "RemoveContainer" containerID="e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.925035 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dnwbs"] Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.944896 4758 scope.go:117] "RemoveContainer" containerID="25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.983453 4758 scope.go:117] "RemoveContainer" containerID="3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba" Oct 04 11:46:47 crc kubenswrapper[4758]: E1004 11:46:47.983911 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba\": container with ID starting with 3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba not found: ID does not exist" containerID="3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.984028 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba"} err="failed to get container status \"3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba\": rpc error: code = NotFound desc = could not find container \"3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba\": container with ID starting with 3a3f688de5d430077b61eb8e8dd72ba5cef7a45338582341be3515324bcf7fba not found: ID does not exist" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.984121 4758 scope.go:117] "RemoveContainer" containerID="e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d" Oct 04 11:46:47 crc kubenswrapper[4758]: E1004 11:46:47.984458 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d\": container with ID starting with e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d not found: ID does not exist" containerID="e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.984538 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d"} err="failed to get container status \"e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d\": rpc error: code = NotFound desc = could not find container \"e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d\": container with ID starting with e5c0a16a472a6c981e4059c7bfc5b1aa3ce989c715cfbfc874a3d9b8ae1c182d not found: ID does not exist" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.984636 4758 scope.go:117] "RemoveContainer" containerID="25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e" Oct 04 11:46:47 crc kubenswrapper[4758]: E1004 11:46:47.984906 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e\": container with ID starting with 25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e not found: ID does not exist" containerID="25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e" Oct 04 11:46:47 crc kubenswrapper[4758]: I1004 11:46:47.984982 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e"} err="failed to get container status \"25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e\": rpc error: code = NotFound desc = could not find container \"25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e\": container with ID starting with 25c6bc5ae7c5d21437423fefabc0997e99599a49660fa829b435cf232a58cd9e not found: ID does not exist" Oct 04 11:46:49 crc kubenswrapper[4758]: I1004 11:46:49.336389 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" path="/var/lib/kubelet/pods/f8ebb362-6343-414e-9369-bb24fa7411a0/volumes" Oct 04 11:46:52 crc kubenswrapper[4758]: I1004 11:46:52.847401 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:52 crc kubenswrapper[4758]: I1004 11:46:52.900088 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:53 crc kubenswrapper[4758]: I1004 11:46:53.664303 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pthxf"] Oct 04 11:46:53 crc kubenswrapper[4758]: I1004 11:46:53.937528 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pthxf" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="registry-server" containerID="cri-o://435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c" gracePeriod=2 Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.388044 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.498071 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-utilities\") pod \"2324546b-1772-42a6-b0ec-87811aebc102\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.498146 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-catalog-content\") pod \"2324546b-1772-42a6-b0ec-87811aebc102\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.498189 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj99d\" (UniqueName: \"kubernetes.io/projected/2324546b-1772-42a6-b0ec-87811aebc102-kube-api-access-fj99d\") pod \"2324546b-1772-42a6-b0ec-87811aebc102\" (UID: \"2324546b-1772-42a6-b0ec-87811aebc102\") " Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.499878 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-utilities" (OuterVolumeSpecName: "utilities") pod "2324546b-1772-42a6-b0ec-87811aebc102" (UID: "2324546b-1772-42a6-b0ec-87811aebc102"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.508346 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2324546b-1772-42a6-b0ec-87811aebc102-kube-api-access-fj99d" (OuterVolumeSpecName: "kube-api-access-fj99d") pod "2324546b-1772-42a6-b0ec-87811aebc102" (UID: "2324546b-1772-42a6-b0ec-87811aebc102"). InnerVolumeSpecName "kube-api-access-fj99d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.584723 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2324546b-1772-42a6-b0ec-87811aebc102" (UID: "2324546b-1772-42a6-b0ec-87811aebc102"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.600392 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.600420 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2324546b-1772-42a6-b0ec-87811aebc102-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.600435 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj99d\" (UniqueName: \"kubernetes.io/projected/2324546b-1772-42a6-b0ec-87811aebc102-kube-api-access-fj99d\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.968607 4758 generic.go:334] "Generic (PLEG): container finished" podID="2324546b-1772-42a6-b0ec-87811aebc102" containerID="435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c" exitCode=0 Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.968649 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerDied","Data":"435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c"} Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.968675 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pthxf" event={"ID":"2324546b-1772-42a6-b0ec-87811aebc102","Type":"ContainerDied","Data":"e6da4ecc89d96f20856c8473d29dd65bae078cb606b2cd2c75bddb35290be60f"} Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.968685 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pthxf" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.968693 4758 scope.go:117] "RemoveContainer" containerID="435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c" Oct 04 11:46:54 crc kubenswrapper[4758]: I1004 11:46:54.994806 4758 scope.go:117] "RemoveContainer" containerID="6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.011235 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pthxf"] Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.020279 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pthxf"] Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.033844 4758 scope.go:117] "RemoveContainer" containerID="ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.064778 4758 scope.go:117] "RemoveContainer" containerID="435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c" Oct 04 11:46:55 crc kubenswrapper[4758]: E1004 11:46:55.068210 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c\": container with ID starting with 435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c not found: ID does not exist" containerID="435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.068240 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c"} err="failed to get container status \"435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c\": rpc error: code = NotFound desc = could not find container \"435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c\": container with ID starting with 435585c921f958c4d6cdc3a17767aae93b3ef5eed31828491db99e1883c1ef1c not found: ID does not exist" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.068260 4758 scope.go:117] "RemoveContainer" containerID="6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d" Oct 04 11:46:55 crc kubenswrapper[4758]: E1004 11:46:55.068574 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d\": container with ID starting with 6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d not found: ID does not exist" containerID="6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.068614 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d"} err="failed to get container status \"6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d\": rpc error: code = NotFound desc = could not find container \"6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d\": container with ID starting with 6af8ebbe67c2cf5540cee53c5dda40a3694925bb8f4f230ec5dcf7317ce34b0d not found: ID does not exist" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.068644 4758 scope.go:117] "RemoveContainer" containerID="ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26" Oct 04 11:46:55 crc kubenswrapper[4758]: E1004 11:46:55.069048 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26\": container with ID starting with ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26 not found: ID does not exist" containerID="ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.069075 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26"} err="failed to get container status \"ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26\": rpc error: code = NotFound desc = could not find container \"ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26\": container with ID starting with ae3cc36c12b27500315e754fa9c3f340456a147b4e4b126ddf2964e03a59ab26 not found: ID does not exist" Oct 04 11:46:55 crc kubenswrapper[4758]: I1004 11:46:55.342664 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2324546b-1772-42a6-b0ec-87811aebc102" path="/var/lib/kubelet/pods/2324546b-1772-42a6-b0ec-87811aebc102/volumes" Oct 04 11:47:01 crc kubenswrapper[4758]: I1004 11:47:01.249323 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:47:01 crc kubenswrapper[4758]: I1004 11:47:01.249824 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:47:31 crc kubenswrapper[4758]: I1004 11:47:31.249056 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:47:31 crc kubenswrapper[4758]: I1004 11:47:31.249899 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.042787 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xnvf7"] Oct 04 11:47:54 crc kubenswrapper[4758]: E1004 11:47:54.043713 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="extract-utilities" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.043725 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="extract-utilities" Oct 04 11:47:54 crc kubenswrapper[4758]: E1004 11:47:54.043738 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="extract-content" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.043744 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="extract-content" Oct 04 11:47:54 crc kubenswrapper[4758]: E1004 11:47:54.043762 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="extract-utilities" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.043768 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="extract-utilities" Oct 04 11:47:54 crc kubenswrapper[4758]: E1004 11:47:54.043781 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="extract-content" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.043786 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="extract-content" Oct 04 11:47:54 crc kubenswrapper[4758]: E1004 11:47:54.043800 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="registry-server" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.043805 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="registry-server" Oct 04 11:47:54 crc kubenswrapper[4758]: E1004 11:47:54.043816 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="registry-server" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.043822 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="registry-server" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.044010 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8ebb362-6343-414e-9369-bb24fa7411a0" containerName="registry-server" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.044024 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="2324546b-1772-42a6-b0ec-87811aebc102" containerName="registry-server" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.045306 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.065462 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xnvf7"] Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.132485 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-utilities\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.132792 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-catalog-content\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.133022 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww6rp\" (UniqueName: \"kubernetes.io/projected/4be2df58-6fea-487b-a91a-101e5015ffed-kube-api-access-ww6rp\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.235069 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww6rp\" (UniqueName: \"kubernetes.io/projected/4be2df58-6fea-487b-a91a-101e5015ffed-kube-api-access-ww6rp\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.235235 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-utilities\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.235269 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-catalog-content\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.235723 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-catalog-content\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.235963 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-utilities\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.254893 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww6rp\" (UniqueName: \"kubernetes.io/projected/4be2df58-6fea-487b-a91a-101e5015ffed-kube-api-access-ww6rp\") pod \"community-operators-xnvf7\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:54 crc kubenswrapper[4758]: I1004 11:47:54.408827 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:47:55 crc kubenswrapper[4758]: I1004 11:47:55.030478 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xnvf7"] Oct 04 11:47:55 crc kubenswrapper[4758]: I1004 11:47:55.502078 4758 generic.go:334] "Generic (PLEG): container finished" podID="4be2df58-6fea-487b-a91a-101e5015ffed" containerID="837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4" exitCode=0 Oct 04 11:47:55 crc kubenswrapper[4758]: I1004 11:47:55.502147 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerDied","Data":"837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4"} Oct 04 11:47:55 crc kubenswrapper[4758]: I1004 11:47:55.502214 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerStarted","Data":"32469ab612652c458a939f4de713605fbc2aebb93b7fe609dfc91cd68d9e05b1"} Oct 04 11:47:56 crc kubenswrapper[4758]: I1004 11:47:56.513501 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerStarted","Data":"b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335"} Oct 04 11:47:58 crc kubenswrapper[4758]: I1004 11:47:58.540864 4758 generic.go:334] "Generic (PLEG): container finished" podID="4be2df58-6fea-487b-a91a-101e5015ffed" containerID="b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335" exitCode=0 Oct 04 11:47:58 crc kubenswrapper[4758]: I1004 11:47:58.541197 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerDied","Data":"b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335"} Oct 04 11:47:59 crc kubenswrapper[4758]: I1004 11:47:59.553218 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerStarted","Data":"574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9"} Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.250014 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.250608 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.250701 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.252149 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bdbd2a04ee42e03967d4eca5f189346fe862da26e4308abe36307465a4dcbbcc"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.252269 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://bdbd2a04ee42e03967d4eca5f189346fe862da26e4308abe36307465a4dcbbcc" gracePeriod=600 Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.573495 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="bdbd2a04ee42e03967d4eca5f189346fe862da26e4308abe36307465a4dcbbcc" exitCode=0 Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.573571 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"bdbd2a04ee42e03967d4eca5f189346fe862da26e4308abe36307465a4dcbbcc"} Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.573937 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7"} Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.573961 4758 scope.go:117] "RemoveContainer" containerID="d3f5ef50ff9a984f7d8d5ebc3e4664ae1ad3ab457fad914fe79f681bd8382a55" Oct 04 11:48:01 crc kubenswrapper[4758]: I1004 11:48:01.597621 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xnvf7" podStartSLOduration=4.109414328 podStartE2EDuration="7.597602422s" podCreationTimestamp="2025-10-04 11:47:54 +0000 UTC" firstStartedPulling="2025-10-04 11:47:55.503795834 +0000 UTC m=+3452.796446723" lastFinishedPulling="2025-10-04 11:47:58.991983928 +0000 UTC m=+3456.284634817" observedRunningTime="2025-10-04 11:47:59.590881446 +0000 UTC m=+3456.883532405" watchObservedRunningTime="2025-10-04 11:48:01.597602422 +0000 UTC m=+3458.890253311" Oct 04 11:48:04 crc kubenswrapper[4758]: I1004 11:48:04.409243 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:48:04 crc kubenswrapper[4758]: I1004 11:48:04.409671 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:48:04 crc kubenswrapper[4758]: I1004 11:48:04.453702 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:48:04 crc kubenswrapper[4758]: I1004 11:48:04.654982 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:48:04 crc kubenswrapper[4758]: I1004 11:48:04.699219 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xnvf7"] Oct 04 11:48:06 crc kubenswrapper[4758]: I1004 11:48:06.618394 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xnvf7" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="registry-server" containerID="cri-o://574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9" gracePeriod=2 Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.117429 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.193003 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-utilities\") pod \"4be2df58-6fea-487b-a91a-101e5015ffed\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.193163 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww6rp\" (UniqueName: \"kubernetes.io/projected/4be2df58-6fea-487b-a91a-101e5015ffed-kube-api-access-ww6rp\") pod \"4be2df58-6fea-487b-a91a-101e5015ffed\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.193190 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-catalog-content\") pod \"4be2df58-6fea-487b-a91a-101e5015ffed\" (UID: \"4be2df58-6fea-487b-a91a-101e5015ffed\") " Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.194237 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-utilities" (OuterVolumeSpecName: "utilities") pod "4be2df58-6fea-487b-a91a-101e5015ffed" (UID: "4be2df58-6fea-487b-a91a-101e5015ffed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.200246 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4be2df58-6fea-487b-a91a-101e5015ffed-kube-api-access-ww6rp" (OuterVolumeSpecName: "kube-api-access-ww6rp") pod "4be2df58-6fea-487b-a91a-101e5015ffed" (UID: "4be2df58-6fea-487b-a91a-101e5015ffed"). InnerVolumeSpecName "kube-api-access-ww6rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.242247 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4be2df58-6fea-487b-a91a-101e5015ffed" (UID: "4be2df58-6fea-487b-a91a-101e5015ffed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.295118 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.295154 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww6rp\" (UniqueName: \"kubernetes.io/projected/4be2df58-6fea-487b-a91a-101e5015ffed-kube-api-access-ww6rp\") on node \"crc\" DevicePath \"\"" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.295163 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4be2df58-6fea-487b-a91a-101e5015ffed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.627569 4758 generic.go:334] "Generic (PLEG): container finished" podID="4be2df58-6fea-487b-a91a-101e5015ffed" containerID="574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9" exitCode=0 Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.627610 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerDied","Data":"574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9"} Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.627622 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnvf7" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.627637 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnvf7" event={"ID":"4be2df58-6fea-487b-a91a-101e5015ffed","Type":"ContainerDied","Data":"32469ab612652c458a939f4de713605fbc2aebb93b7fe609dfc91cd68d9e05b1"} Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.627670 4758 scope.go:117] "RemoveContainer" containerID="574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.653546 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xnvf7"] Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.658374 4758 scope.go:117] "RemoveContainer" containerID="b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.663915 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xnvf7"] Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.682267 4758 scope.go:117] "RemoveContainer" containerID="837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.721639 4758 scope.go:117] "RemoveContainer" containerID="574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9" Oct 04 11:48:07 crc kubenswrapper[4758]: E1004 11:48:07.722011 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9\": container with ID starting with 574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9 not found: ID does not exist" containerID="574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.722067 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9"} err="failed to get container status \"574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9\": rpc error: code = NotFound desc = could not find container \"574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9\": container with ID starting with 574351dc122ce3e0d4d37e5eb25f70ce9d29cd4111ce26864415cdf53b867bb9 not found: ID does not exist" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.722125 4758 scope.go:117] "RemoveContainer" containerID="b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335" Oct 04 11:48:07 crc kubenswrapper[4758]: E1004 11:48:07.722892 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335\": container with ID starting with b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335 not found: ID does not exist" containerID="b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.722936 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335"} err="failed to get container status \"b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335\": rpc error: code = NotFound desc = could not find container \"b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335\": container with ID starting with b3204e9b66da46f71abc369c976ff23593b5b4694f76ce56de201a6e6d343335 not found: ID does not exist" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.722954 4758 scope.go:117] "RemoveContainer" containerID="837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4" Oct 04 11:48:07 crc kubenswrapper[4758]: E1004 11:48:07.723791 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4\": container with ID starting with 837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4 not found: ID does not exist" containerID="837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4" Oct 04 11:48:07 crc kubenswrapper[4758]: I1004 11:48:07.723831 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4"} err="failed to get container status \"837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4\": rpc error: code = NotFound desc = could not find container \"837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4\": container with ID starting with 837f916d40ff4a5a1eed5d9dffdd91b0cd84e3aa04006891f2d1706c239ff0b4 not found: ID does not exist" Oct 04 11:48:09 crc kubenswrapper[4758]: I1004 11:48:09.336423 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" path="/var/lib/kubelet/pods/4be2df58-6fea-487b-a91a-101e5015ffed/volumes" Oct 04 11:50:01 crc kubenswrapper[4758]: I1004 11:50:01.249292 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:50:01 crc kubenswrapper[4758]: I1004 11:50:01.249977 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:50:31 crc kubenswrapper[4758]: I1004 11:50:31.249335 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:50:31 crc kubenswrapper[4758]: I1004 11:50:31.249850 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.597285 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vf6gl"] Oct 04 11:50:52 crc kubenswrapper[4758]: E1004 11:50:52.598452 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="extract-utilities" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.598467 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="extract-utilities" Oct 04 11:50:52 crc kubenswrapper[4758]: E1004 11:50:52.598480 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="registry-server" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.598489 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="registry-server" Oct 04 11:50:52 crc kubenswrapper[4758]: E1004 11:50:52.598516 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="extract-content" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.598525 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="extract-content" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.598777 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="4be2df58-6fea-487b-a91a-101e5015ffed" containerName="registry-server" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.600400 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.654625 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vf6gl"] Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.774470 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2g6s\" (UniqueName: \"kubernetes.io/projected/f3a3525b-d316-4430-a28f-02726be9c506-kube-api-access-p2g6s\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.774525 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-catalog-content\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.774622 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-utilities\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.876086 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-utilities\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.876888 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2g6s\" (UniqueName: \"kubernetes.io/projected/f3a3525b-d316-4430-a28f-02726be9c506-kube-api-access-p2g6s\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.877335 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-catalog-content\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.877720 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-catalog-content\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.876778 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-utilities\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.900373 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2g6s\" (UniqueName: \"kubernetes.io/projected/f3a3525b-d316-4430-a28f-02726be9c506-kube-api-access-p2g6s\") pod \"redhat-marketplace-vf6gl\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:52 crc kubenswrapper[4758]: I1004 11:50:52.921152 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:50:53 crc kubenswrapper[4758]: I1004 11:50:53.470042 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vf6gl"] Oct 04 11:50:54 crc kubenswrapper[4758]: I1004 11:50:54.033509 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3a3525b-d316-4430-a28f-02726be9c506" containerID="cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4" exitCode=0 Oct 04 11:50:54 crc kubenswrapper[4758]: I1004 11:50:54.033597 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerDied","Data":"cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4"} Oct 04 11:50:54 crc kubenswrapper[4758]: I1004 11:50:54.033833 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerStarted","Data":"8b5b6d5756826af34e45359af113083a66a80b135ecbe3777bc9640d3e624183"} Oct 04 11:50:55 crc kubenswrapper[4758]: I1004 11:50:55.054434 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerStarted","Data":"f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a"} Oct 04 11:50:56 crc kubenswrapper[4758]: I1004 11:50:56.066697 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3a3525b-d316-4430-a28f-02726be9c506" containerID="f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a" exitCode=0 Oct 04 11:50:56 crc kubenswrapper[4758]: I1004 11:50:56.066767 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerDied","Data":"f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a"} Oct 04 11:50:57 crc kubenswrapper[4758]: I1004 11:50:57.078589 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerStarted","Data":"949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5"} Oct 04 11:50:57 crc kubenswrapper[4758]: I1004 11:50:57.097020 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vf6gl" podStartSLOduration=2.582402739 podStartE2EDuration="5.097001214s" podCreationTimestamp="2025-10-04 11:50:52 +0000 UTC" firstStartedPulling="2025-10-04 11:50:54.035567185 +0000 UTC m=+3631.328218074" lastFinishedPulling="2025-10-04 11:50:56.55016566 +0000 UTC m=+3633.842816549" observedRunningTime="2025-10-04 11:50:57.094015294 +0000 UTC m=+3634.386666203" watchObservedRunningTime="2025-10-04 11:50:57.097001214 +0000 UTC m=+3634.389652123" Oct 04 11:51:01 crc kubenswrapper[4758]: I1004 11:51:01.249899 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:51:01 crc kubenswrapper[4758]: I1004 11:51:01.250287 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:51:01 crc kubenswrapper[4758]: I1004 11:51:01.250349 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:51:01 crc kubenswrapper[4758]: I1004 11:51:01.251482 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:51:01 crc kubenswrapper[4758]: I1004 11:51:01.251586 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" gracePeriod=600 Oct 04 11:51:01 crc kubenswrapper[4758]: E1004 11:51:01.379053 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.126168 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" exitCode=0 Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.126220 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7"} Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.126265 4758 scope.go:117] "RemoveContainer" containerID="bdbd2a04ee42e03967d4eca5f189346fe862da26e4308abe36307465a4dcbbcc" Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.126849 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:51:02 crc kubenswrapper[4758]: E1004 11:51:02.127208 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.921746 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.922068 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:51:02 crc kubenswrapper[4758]: I1004 11:51:02.970437 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:51:03 crc kubenswrapper[4758]: I1004 11:51:03.190265 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:51:03 crc kubenswrapper[4758]: I1004 11:51:03.234665 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vf6gl"] Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.157925 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vf6gl" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="registry-server" containerID="cri-o://949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5" gracePeriod=2 Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.607194 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.712490 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2g6s\" (UniqueName: \"kubernetes.io/projected/f3a3525b-d316-4430-a28f-02726be9c506-kube-api-access-p2g6s\") pod \"f3a3525b-d316-4430-a28f-02726be9c506\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.713000 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-utilities\") pod \"f3a3525b-d316-4430-a28f-02726be9c506\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.713028 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-catalog-content\") pod \"f3a3525b-d316-4430-a28f-02726be9c506\" (UID: \"f3a3525b-d316-4430-a28f-02726be9c506\") " Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.713718 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-utilities" (OuterVolumeSpecName: "utilities") pod "f3a3525b-d316-4430-a28f-02726be9c506" (UID: "f3a3525b-d316-4430-a28f-02726be9c506"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.718246 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3a3525b-d316-4430-a28f-02726be9c506-kube-api-access-p2g6s" (OuterVolumeSpecName: "kube-api-access-p2g6s") pod "f3a3525b-d316-4430-a28f-02726be9c506" (UID: "f3a3525b-d316-4430-a28f-02726be9c506"). InnerVolumeSpecName "kube-api-access-p2g6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.727187 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3a3525b-d316-4430-a28f-02726be9c506" (UID: "f3a3525b-d316-4430-a28f-02726be9c506"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.815156 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.815184 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3a3525b-d316-4430-a28f-02726be9c506-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:51:05 crc kubenswrapper[4758]: I1004 11:51:05.815195 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2g6s\" (UniqueName: \"kubernetes.io/projected/f3a3525b-d316-4430-a28f-02726be9c506-kube-api-access-p2g6s\") on node \"crc\" DevicePath \"\"" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.169612 4758 generic.go:334] "Generic (PLEG): container finished" podID="f3a3525b-d316-4430-a28f-02726be9c506" containerID="949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5" exitCode=0 Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.169651 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerDied","Data":"949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5"} Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.169677 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vf6gl" event={"ID":"f3a3525b-d316-4430-a28f-02726be9c506","Type":"ContainerDied","Data":"8b5b6d5756826af34e45359af113083a66a80b135ecbe3777bc9640d3e624183"} Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.169694 4758 scope.go:117] "RemoveContainer" containerID="949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.171364 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vf6gl" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.197723 4758 scope.go:117] "RemoveContainer" containerID="f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.233969 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vf6gl"] Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.236594 4758 scope.go:117] "RemoveContainer" containerID="cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.247134 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vf6gl"] Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.266779 4758 scope.go:117] "RemoveContainer" containerID="949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5" Oct 04 11:51:06 crc kubenswrapper[4758]: E1004 11:51:06.274276 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5\": container with ID starting with 949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5 not found: ID does not exist" containerID="949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.274331 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5"} err="failed to get container status \"949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5\": rpc error: code = NotFound desc = could not find container \"949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5\": container with ID starting with 949b892d2c2b31e1507ef02a9610cf0605fa10460d0575f6f5812449efb26cd5 not found: ID does not exist" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.274358 4758 scope.go:117] "RemoveContainer" containerID="f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a" Oct 04 11:51:06 crc kubenswrapper[4758]: E1004 11:51:06.274893 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a\": container with ID starting with f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a not found: ID does not exist" containerID="f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.274919 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a"} err="failed to get container status \"f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a\": rpc error: code = NotFound desc = could not find container \"f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a\": container with ID starting with f124f053d4daee0778f6a57cf11e70764b0b935e9447e8168f904a53bfbfda1a not found: ID does not exist" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.274957 4758 scope.go:117] "RemoveContainer" containerID="cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4" Oct 04 11:51:06 crc kubenswrapper[4758]: E1004 11:51:06.275345 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4\": container with ID starting with cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4 not found: ID does not exist" containerID="cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4" Oct 04 11:51:06 crc kubenswrapper[4758]: I1004 11:51:06.275382 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4"} err="failed to get container status \"cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4\": rpc error: code = NotFound desc = could not find container \"cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4\": container with ID starting with cca7d2f65e0dfa9bfd408970b094f719b272532e1c1bd8f805c3b337316904f4 not found: ID does not exist" Oct 04 11:51:07 crc kubenswrapper[4758]: I1004 11:51:07.335194 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3a3525b-d316-4430-a28f-02726be9c506" path="/var/lib/kubelet/pods/f3a3525b-d316-4430-a28f-02726be9c506/volumes" Oct 04 11:51:17 crc kubenswrapper[4758]: I1004 11:51:17.326505 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:51:17 crc kubenswrapper[4758]: E1004 11:51:17.327183 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:51:30 crc kubenswrapper[4758]: I1004 11:51:30.325650 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:51:30 crc kubenswrapper[4758]: E1004 11:51:30.326319 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:51:41 crc kubenswrapper[4758]: I1004 11:51:41.326014 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:51:41 crc kubenswrapper[4758]: E1004 11:51:41.326729 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:51:54 crc kubenswrapper[4758]: I1004 11:51:54.326434 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:51:54 crc kubenswrapper[4758]: E1004 11:51:54.327197 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:52:09 crc kubenswrapper[4758]: I1004 11:52:09.328345 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:52:09 crc kubenswrapper[4758]: E1004 11:52:09.329112 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:52:20 crc kubenswrapper[4758]: I1004 11:52:20.325899 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:52:20 crc kubenswrapper[4758]: E1004 11:52:20.326839 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:52:35 crc kubenswrapper[4758]: I1004 11:52:35.327159 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:52:35 crc kubenswrapper[4758]: E1004 11:52:35.328088 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:52:47 crc kubenswrapper[4758]: I1004 11:52:47.325695 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:52:47 crc kubenswrapper[4758]: E1004 11:52:47.326472 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:52:58 crc kubenswrapper[4758]: I1004 11:52:58.325850 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:52:58 crc kubenswrapper[4758]: E1004 11:52:58.326679 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:53:13 crc kubenswrapper[4758]: I1004 11:53:13.337446 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:53:13 crc kubenswrapper[4758]: E1004 11:53:13.339359 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:53:27 crc kubenswrapper[4758]: I1004 11:53:27.325869 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:53:27 crc kubenswrapper[4758]: E1004 11:53:27.327576 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:53:38 crc kubenswrapper[4758]: I1004 11:53:38.326722 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:53:38 crc kubenswrapper[4758]: E1004 11:53:38.328600 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:53:49 crc kubenswrapper[4758]: I1004 11:53:49.326117 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:53:49 crc kubenswrapper[4758]: E1004 11:53:49.327289 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:54:00 crc kubenswrapper[4758]: I1004 11:54:00.326558 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:54:00 crc kubenswrapper[4758]: E1004 11:54:00.327345 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:54:13 crc kubenswrapper[4758]: I1004 11:54:13.335890 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:54:13 crc kubenswrapper[4758]: E1004 11:54:13.336746 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:54:28 crc kubenswrapper[4758]: I1004 11:54:28.331097 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:54:28 crc kubenswrapper[4758]: E1004 11:54:28.332186 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:54:42 crc kubenswrapper[4758]: I1004 11:54:42.327983 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:54:42 crc kubenswrapper[4758]: E1004 11:54:42.328950 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:54:54 crc kubenswrapper[4758]: I1004 11:54:54.325570 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:54:54 crc kubenswrapper[4758]: E1004 11:54:54.327014 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:55:06 crc kubenswrapper[4758]: I1004 11:55:06.327069 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:55:06 crc kubenswrapper[4758]: E1004 11:55:06.327715 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:55:18 crc kubenswrapper[4758]: I1004 11:55:18.325504 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:55:18 crc kubenswrapper[4758]: E1004 11:55:18.326324 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:55:31 crc kubenswrapper[4758]: I1004 11:55:31.325625 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:55:31 crc kubenswrapper[4758]: E1004 11:55:31.326470 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:55:44 crc kubenswrapper[4758]: I1004 11:55:44.327184 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:55:44 crc kubenswrapper[4758]: E1004 11:55:44.329798 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:55:59 crc kubenswrapper[4758]: I1004 11:55:59.327589 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:55:59 crc kubenswrapper[4758]: E1004 11:55:59.334517 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 11:56:14 crc kubenswrapper[4758]: I1004 11:56:14.325604 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:56:14 crc kubenswrapper[4758]: I1004 11:56:14.907555 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"2320c5922cab48a5f665e70e3a26f8604ef267fa9be39ebc5e9e9f7ea77c98b3"} Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.609372 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-44prw"] Oct 04 11:57:10 crc kubenswrapper[4758]: E1004 11:57:10.610286 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="extract-content" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.610298 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="extract-content" Oct 04 11:57:10 crc kubenswrapper[4758]: E1004 11:57:10.610327 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="extract-utilities" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.610334 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="extract-utilities" Oct 04 11:57:10 crc kubenswrapper[4758]: E1004 11:57:10.610345 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="registry-server" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.610353 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="registry-server" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.610549 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a3525b-d316-4430-a28f-02726be9c506" containerName="registry-server" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.611802 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.625399 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44prw"] Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.686233 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nswd\" (UniqueName: \"kubernetes.io/projected/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-kube-api-access-4nswd\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.686327 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-utilities\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.686355 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-catalog-content\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.787629 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-utilities\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.787672 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-catalog-content\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.787784 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nswd\" (UniqueName: \"kubernetes.io/projected/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-kube-api-access-4nswd\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.788570 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-utilities\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.788629 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-catalog-content\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.814760 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nswd\" (UniqueName: \"kubernetes.io/projected/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-kube-api-access-4nswd\") pod \"redhat-operators-44prw\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:10 crc kubenswrapper[4758]: I1004 11:57:10.944038 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:11 crc kubenswrapper[4758]: I1004 11:57:11.416054 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44prw"] Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.424319 4758 generic.go:334] "Generic (PLEG): container finished" podID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerID="0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9" exitCode=0 Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.424378 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerDied","Data":"0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9"} Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.424855 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerStarted","Data":"0e0b6a535d8912ca7fe27f16ce12dc9bf9a721de65541b6d58ed38a83ec8451b"} Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.427494 4758 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.808250 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jfw5k"] Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.812754 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.823540 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jfw5k"] Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.925470 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-catalog-content\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.925795 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-utilities\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:12 crc kubenswrapper[4758]: I1004 11:57:12.925989 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb8tc\" (UniqueName: \"kubernetes.io/projected/d807f53b-de81-4144-8d40-83a550ff9c60-kube-api-access-xb8tc\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.027668 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-utilities\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.027757 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb8tc\" (UniqueName: \"kubernetes.io/projected/d807f53b-de81-4144-8d40-83a550ff9c60-kube-api-access-xb8tc\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.027841 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-catalog-content\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.028412 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-catalog-content\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.028477 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-utilities\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.052480 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb8tc\" (UniqueName: \"kubernetes.io/projected/d807f53b-de81-4144-8d40-83a550ff9c60-kube-api-access-xb8tc\") pod \"certified-operators-jfw5k\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.223576 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:13 crc kubenswrapper[4758]: I1004 11:57:13.559618 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jfw5k"] Oct 04 11:57:14 crc kubenswrapper[4758]: I1004 11:57:14.446004 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerStarted","Data":"6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173"} Oct 04 11:57:14 crc kubenswrapper[4758]: I1004 11:57:14.446313 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerStarted","Data":"ab448db13d9fb6f7b6bf47ea4510409ed4e46c6719f089eab069670acb9e347c"} Oct 04 11:57:14 crc kubenswrapper[4758]: I1004 11:57:14.448243 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerStarted","Data":"6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21"} Oct 04 11:57:15 crc kubenswrapper[4758]: I1004 11:57:15.455756 4758 generic.go:334] "Generic (PLEG): container finished" podID="d807f53b-de81-4144-8d40-83a550ff9c60" containerID="6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173" exitCode=0 Oct 04 11:57:15 crc kubenswrapper[4758]: I1004 11:57:15.455971 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerDied","Data":"6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173"} Oct 04 11:57:16 crc kubenswrapper[4758]: I1004 11:57:16.466226 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerStarted","Data":"47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed"} Oct 04 11:57:18 crc kubenswrapper[4758]: I1004 11:57:18.484984 4758 generic.go:334] "Generic (PLEG): container finished" podID="d807f53b-de81-4144-8d40-83a550ff9c60" containerID="47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed" exitCode=0 Oct 04 11:57:18 crc kubenswrapper[4758]: I1004 11:57:18.485089 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerDied","Data":"47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed"} Oct 04 11:57:18 crc kubenswrapper[4758]: I1004 11:57:18.488445 4758 generic.go:334] "Generic (PLEG): container finished" podID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerID="6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21" exitCode=0 Oct 04 11:57:18 crc kubenswrapper[4758]: I1004 11:57:18.488514 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerDied","Data":"6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21"} Oct 04 11:57:19 crc kubenswrapper[4758]: I1004 11:57:19.498451 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerStarted","Data":"24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515"} Oct 04 11:57:19 crc kubenswrapper[4758]: I1004 11:57:19.500738 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerStarted","Data":"b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8"} Oct 04 11:57:19 crc kubenswrapper[4758]: I1004 11:57:19.527460 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jfw5k" podStartSLOduration=4.073609842 podStartE2EDuration="7.527441536s" podCreationTimestamp="2025-10-04 11:57:12 +0000 UTC" firstStartedPulling="2025-10-04 11:57:15.458798189 +0000 UTC m=+4012.751449078" lastFinishedPulling="2025-10-04 11:57:18.912629883 +0000 UTC m=+4016.205280772" observedRunningTime="2025-10-04 11:57:19.519936875 +0000 UTC m=+4016.812587764" watchObservedRunningTime="2025-10-04 11:57:19.527441536 +0000 UTC m=+4016.820092425" Oct 04 11:57:19 crc kubenswrapper[4758]: I1004 11:57:19.541026 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-44prw" podStartSLOduration=3.082328362 podStartE2EDuration="9.541009741s" podCreationTimestamp="2025-10-04 11:57:10 +0000 UTC" firstStartedPulling="2025-10-04 11:57:12.427279938 +0000 UTC m=+4009.719930827" lastFinishedPulling="2025-10-04 11:57:18.885961317 +0000 UTC m=+4016.178612206" observedRunningTime="2025-10-04 11:57:19.537054285 +0000 UTC m=+4016.829705174" watchObservedRunningTime="2025-10-04 11:57:19.541009741 +0000 UTC m=+4016.833660630" Oct 04 11:57:20 crc kubenswrapper[4758]: I1004 11:57:20.944214 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:20 crc kubenswrapper[4758]: I1004 11:57:20.944575 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:21 crc kubenswrapper[4758]: I1004 11:57:21.987531 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-44prw" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="registry-server" probeResult="failure" output=< Oct 04 11:57:21 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:57:21 crc kubenswrapper[4758]: > Oct 04 11:57:22 crc kubenswrapper[4758]: I1004 11:57:22.528795 4758 generic.go:334] "Generic (PLEG): container finished" podID="8a041496-5b8c-4d15-961a-d710b5a461fe" containerID="db31500cd99d82268b9b85556f8fd21ca59e05dcfca81fd0d345c081052a530f" exitCode=1 Oct 04 11:57:22 crc kubenswrapper[4758]: I1004 11:57:22.528891 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8a041496-5b8c-4d15-961a-d710b5a461fe","Type":"ContainerDied","Data":"db31500cd99d82268b9b85556f8fd21ca59e05dcfca81fd0d345c081052a530f"} Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.224179 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.224252 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.415489 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.838179 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.933880 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ca-certs\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.933992 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-config-data\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934039 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config-secret\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934081 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-temporary\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934123 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-workdir\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934224 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ssh-key\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934261 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934287 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.934333 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ddl8\" (UniqueName: \"kubernetes.io/projected/8a041496-5b8c-4d15-961a-d710b5a461fe-kube-api-access-5ddl8\") pod \"8a041496-5b8c-4d15-961a-d710b5a461fe\" (UID: \"8a041496-5b8c-4d15-961a-d710b5a461fe\") " Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.935515 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-config-data" (OuterVolumeSpecName: "config-data") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.936437 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.941387 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a041496-5b8c-4d15-961a-d710b5a461fe-kube-api-access-5ddl8" (OuterVolumeSpecName: "kube-api-access-5ddl8") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "kube-api-access-5ddl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.947513 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.958415 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.972782 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.976568 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:57:23 crc kubenswrapper[4758]: I1004 11:57:23.985792 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.004799 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8a041496-5b8c-4d15-961a-d710b5a461fe" (UID: "8a041496-5b8c-4d15-961a-d710b5a461fe"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036035 4758 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036094 4758 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036122 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036134 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ddl8\" (UniqueName: \"kubernetes.io/projected/8a041496-5b8c-4d15-961a-d710b5a461fe-kube-api-access-5ddl8\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036143 4758 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036150 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a041496-5b8c-4d15-961a-d710b5a461fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036162 4758 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a041496-5b8c-4d15-961a-d710b5a461fe-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036170 4758 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.036179 4758 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8a041496-5b8c-4d15-961a-d710b5a461fe-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.058801 4758 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.138085 4758 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.582712 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"8a041496-5b8c-4d15-961a-d710b5a461fe","Type":"ContainerDied","Data":"2506b20b0aab557c3a1a225dade4a5bfe186e9deba016fe5d809372c1632d557"} Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.582768 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2506b20b0aab557c3a1a225dade4a5bfe186e9deba016fe5d809372c1632d557" Oct 04 11:57:24 crc kubenswrapper[4758]: I1004 11:57:24.583020 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.431913 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 11:57:28 crc kubenswrapper[4758]: E1004 11:57:28.435494 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a041496-5b8c-4d15-961a-d710b5a461fe" containerName="tempest-tests-tempest-tests-runner" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.435855 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a041496-5b8c-4d15-961a-d710b5a461fe" containerName="tempest-tests-tempest-tests-runner" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.436539 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a041496-5b8c-4d15-961a-d710b5a461fe" containerName="tempest-tests-tempest-tests-runner" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.440805 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.442166 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.445274 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9jwgs" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.538688 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh5gp\" (UniqueName: \"kubernetes.io/projected/2c5474d6-89c4-4089-9681-4be34b6964aa-kube-api-access-dh5gp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.538862 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.641566 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh5gp\" (UniqueName: \"kubernetes.io/projected/2c5474d6-89c4-4089-9681-4be34b6964aa-kube-api-access-dh5gp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.641682 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.642378 4758 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.681741 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh5gp\" (UniqueName: \"kubernetes.io/projected/2c5474d6-89c4-4089-9681-4be34b6964aa-kube-api-access-dh5gp\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.686180 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2c5474d6-89c4-4089-9681-4be34b6964aa\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:28 crc kubenswrapper[4758]: I1004 11:57:28.780318 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:57:29 crc kubenswrapper[4758]: I1004 11:57:29.267863 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 11:57:29 crc kubenswrapper[4758]: I1004 11:57:29.634114 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2c5474d6-89c4-4089-9681-4be34b6964aa","Type":"ContainerStarted","Data":"24dea139b8dbc1ec3d14d892f9d445d6b50dd4c8da36b2a2a59ecaf434908ac7"} Oct 04 11:57:30 crc kubenswrapper[4758]: I1004 11:57:30.644237 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2c5474d6-89c4-4089-9681-4be34b6964aa","Type":"ContainerStarted","Data":"abe7de5a0d5aeffca85a0d797942834b4cf880df8ca2fb273e82b63b2365644d"} Oct 04 11:57:32 crc kubenswrapper[4758]: I1004 11:57:32.005840 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-44prw" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="registry-server" probeResult="failure" output=< Oct 04 11:57:32 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:57:32 crc kubenswrapper[4758]: > Oct 04 11:57:33 crc kubenswrapper[4758]: I1004 11:57:33.731399 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:33 crc kubenswrapper[4758]: I1004 11:57:33.759996 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=4.814902206 podStartE2EDuration="5.759974109s" podCreationTimestamp="2025-10-04 11:57:28 +0000 UTC" firstStartedPulling="2025-10-04 11:57:29.269914863 +0000 UTC m=+4026.562565762" lastFinishedPulling="2025-10-04 11:57:30.214986756 +0000 UTC m=+4027.507637665" observedRunningTime="2025-10-04 11:57:30.657620955 +0000 UTC m=+4027.950271844" watchObservedRunningTime="2025-10-04 11:57:33.759974109 +0000 UTC m=+4031.052625018" Oct 04 11:57:33 crc kubenswrapper[4758]: I1004 11:57:33.787426 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jfw5k"] Oct 04 11:57:34 crc kubenswrapper[4758]: I1004 11:57:34.689421 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jfw5k" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="registry-server" containerID="cri-o://24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515" gracePeriod=2 Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.111771 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.167783 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-utilities\") pod \"d807f53b-de81-4144-8d40-83a550ff9c60\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.167828 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-catalog-content\") pod \"d807f53b-de81-4144-8d40-83a550ff9c60\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.167990 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb8tc\" (UniqueName: \"kubernetes.io/projected/d807f53b-de81-4144-8d40-83a550ff9c60-kube-api-access-xb8tc\") pod \"d807f53b-de81-4144-8d40-83a550ff9c60\" (UID: \"d807f53b-de81-4144-8d40-83a550ff9c60\") " Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.168620 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-utilities" (OuterVolumeSpecName: "utilities") pod "d807f53b-de81-4144-8d40-83a550ff9c60" (UID: "d807f53b-de81-4144-8d40-83a550ff9c60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.176307 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d807f53b-de81-4144-8d40-83a550ff9c60-kube-api-access-xb8tc" (OuterVolumeSpecName: "kube-api-access-xb8tc") pod "d807f53b-de81-4144-8d40-83a550ff9c60" (UID: "d807f53b-de81-4144-8d40-83a550ff9c60"). InnerVolumeSpecName "kube-api-access-xb8tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.211129 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d807f53b-de81-4144-8d40-83a550ff9c60" (UID: "d807f53b-de81-4144-8d40-83a550ff9c60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.269950 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.269985 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d807f53b-de81-4144-8d40-83a550ff9c60-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.270000 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb8tc\" (UniqueName: \"kubernetes.io/projected/d807f53b-de81-4144-8d40-83a550ff9c60-kube-api-access-xb8tc\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.704092 4758 generic.go:334] "Generic (PLEG): container finished" podID="d807f53b-de81-4144-8d40-83a550ff9c60" containerID="24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515" exitCode=0 Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.704164 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerDied","Data":"24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515"} Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.704238 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfw5k" event={"ID":"d807f53b-de81-4144-8d40-83a550ff9c60","Type":"ContainerDied","Data":"ab448db13d9fb6f7b6bf47ea4510409ed4e46c6719f089eab069670acb9e347c"} Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.704259 4758 scope.go:117] "RemoveContainer" containerID="24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.704278 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jfw5k" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.733658 4758 scope.go:117] "RemoveContainer" containerID="47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.740416 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jfw5k"] Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.755327 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jfw5k"] Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.765848 4758 scope.go:117] "RemoveContainer" containerID="6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.793599 4758 scope.go:117] "RemoveContainer" containerID="24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515" Oct 04 11:57:35 crc kubenswrapper[4758]: E1004 11:57:35.794358 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515\": container with ID starting with 24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515 not found: ID does not exist" containerID="24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.794409 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515"} err="failed to get container status \"24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515\": rpc error: code = NotFound desc = could not find container \"24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515\": container with ID starting with 24b8ab9b667384aafa7e781efc1a5eab2c23ce2027ae1fbb5b7565e6b518b515 not found: ID does not exist" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.794446 4758 scope.go:117] "RemoveContainer" containerID="47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed" Oct 04 11:57:35 crc kubenswrapper[4758]: E1004 11:57:35.795002 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed\": container with ID starting with 47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed not found: ID does not exist" containerID="47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.795057 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed"} err="failed to get container status \"47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed\": rpc error: code = NotFound desc = could not find container \"47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed\": container with ID starting with 47d2fe4958598149bf6ed74c26768aecb300599fff65a9f5c0b8a38301ea33ed not found: ID does not exist" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.795094 4758 scope.go:117] "RemoveContainer" containerID="6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173" Oct 04 11:57:35 crc kubenswrapper[4758]: E1004 11:57:35.795493 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173\": container with ID starting with 6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173 not found: ID does not exist" containerID="6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173" Oct 04 11:57:35 crc kubenswrapper[4758]: I1004 11:57:35.795527 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173"} err="failed to get container status \"6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173\": rpc error: code = NotFound desc = could not find container \"6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173\": container with ID starting with 6a261eda9f33aa5adca094308b84a9c6a269182c0e379b34b8066d11cff28173 not found: ID does not exist" Oct 04 11:57:37 crc kubenswrapper[4758]: I1004 11:57:37.335631 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" path="/var/lib/kubelet/pods/d807f53b-de81-4144-8d40-83a550ff9c60/volumes" Oct 04 11:57:41 crc kubenswrapper[4758]: I1004 11:57:41.010702 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:41 crc kubenswrapper[4758]: I1004 11:57:41.080937 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:41 crc kubenswrapper[4758]: I1004 11:57:41.834200 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44prw"] Oct 04 11:57:42 crc kubenswrapper[4758]: I1004 11:57:42.776314 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-44prw" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="registry-server" containerID="cri-o://b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8" gracePeriod=2 Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.254515 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.344833 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-utilities\") pod \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.344912 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-catalog-content\") pod \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.345089 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nswd\" (UniqueName: \"kubernetes.io/projected/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-kube-api-access-4nswd\") pod \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\" (UID: \"d9a1c836-bdbf-48dd-9c79-2e7a50c61999\") " Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.349896 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-utilities" (OuterVolumeSpecName: "utilities") pod "d9a1c836-bdbf-48dd-9c79-2e7a50c61999" (UID: "d9a1c836-bdbf-48dd-9c79-2e7a50c61999"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.391965 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-kube-api-access-4nswd" (OuterVolumeSpecName: "kube-api-access-4nswd") pod "d9a1c836-bdbf-48dd-9c79-2e7a50c61999" (UID: "d9a1c836-bdbf-48dd-9c79-2e7a50c61999"). InnerVolumeSpecName "kube-api-access-4nswd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.440636 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9a1c836-bdbf-48dd-9c79-2e7a50c61999" (UID: "d9a1c836-bdbf-48dd-9c79-2e7a50c61999"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.447936 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.447963 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.447976 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nswd\" (UniqueName: \"kubernetes.io/projected/d9a1c836-bdbf-48dd-9c79-2e7a50c61999-kube-api-access-4nswd\") on node \"crc\" DevicePath \"\"" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.793397 4758 generic.go:334] "Generic (PLEG): container finished" podID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerID="b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8" exitCode=0 Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.793766 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerDied","Data":"b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8"} Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.793809 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44prw" event={"ID":"d9a1c836-bdbf-48dd-9c79-2e7a50c61999","Type":"ContainerDied","Data":"0e0b6a535d8912ca7fe27f16ce12dc9bf9a721de65541b6d58ed38a83ec8451b"} Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.793844 4758 scope.go:117] "RemoveContainer" containerID="b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.794362 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44prw" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.844258 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44prw"] Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.852026 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-44prw"] Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.853380 4758 scope.go:117] "RemoveContainer" containerID="6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.878779 4758 scope.go:117] "RemoveContainer" containerID="0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.930487 4758 scope.go:117] "RemoveContainer" containerID="b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8" Oct 04 11:57:43 crc kubenswrapper[4758]: E1004 11:57:43.930867 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8\": container with ID starting with b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8 not found: ID does not exist" containerID="b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.930896 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8"} err="failed to get container status \"b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8\": rpc error: code = NotFound desc = could not find container \"b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8\": container with ID starting with b5579821c85499c8435280c4624ebe7f626e432768b9cc2fd2a5cf7e45c363b8 not found: ID does not exist" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.930918 4758 scope.go:117] "RemoveContainer" containerID="6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21" Oct 04 11:57:43 crc kubenswrapper[4758]: E1004 11:57:43.931127 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21\": container with ID starting with 6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21 not found: ID does not exist" containerID="6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.931146 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21"} err="failed to get container status \"6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21\": rpc error: code = NotFound desc = could not find container \"6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21\": container with ID starting with 6a65346452426f8675c10c1f44adfc6d79e99216fa27fd34a592c31ac0e6fc21 not found: ID does not exist" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.931163 4758 scope.go:117] "RemoveContainer" containerID="0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9" Oct 04 11:57:43 crc kubenswrapper[4758]: E1004 11:57:43.932428 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9\": container with ID starting with 0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9 not found: ID does not exist" containerID="0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9" Oct 04 11:57:43 crc kubenswrapper[4758]: I1004 11:57:43.932472 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9"} err="failed to get container status \"0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9\": rpc error: code = NotFound desc = could not find container \"0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9\": container with ID starting with 0c459e34f0a3cf80b37a0285af10f438f9ec763fb43ade46b52185767f2c8fe9 not found: ID does not exist" Oct 04 11:57:45 crc kubenswrapper[4758]: I1004 11:57:45.337951 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" path="/var/lib/kubelet/pods/d9a1c836-bdbf-48dd-9c79-2e7a50c61999/volumes" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.112781 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hm8tb/must-gather-mshtd"] Oct 04 11:58:06 crc kubenswrapper[4758]: E1004 11:58:06.113764 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="extract-utilities" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.113776 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="extract-utilities" Oct 04 11:58:06 crc kubenswrapper[4758]: E1004 11:58:06.113794 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="extract-content" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.113808 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="extract-content" Oct 04 11:58:06 crc kubenswrapper[4758]: E1004 11:58:06.113821 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="extract-utilities" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.113829 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="extract-utilities" Oct 04 11:58:06 crc kubenswrapper[4758]: E1004 11:58:06.113843 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="registry-server" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.113850 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="registry-server" Oct 04 11:58:06 crc kubenswrapper[4758]: E1004 11:58:06.113884 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="extract-content" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.113891 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="extract-content" Oct 04 11:58:06 crc kubenswrapper[4758]: E1004 11:58:06.113902 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="registry-server" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.113907 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="registry-server" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.114076 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d807f53b-de81-4144-8d40-83a550ff9c60" containerName="registry-server" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.114089 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9a1c836-bdbf-48dd-9c79-2e7a50c61999" containerName="registry-server" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.115034 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.119927 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hm8tb"/"openshift-service-ca.crt" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.120760 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hm8tb/must-gather-mshtd"] Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.123078 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hm8tb"/"default-dockercfg-5ptgz" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.123388 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hm8tb"/"kube-root-ca.crt" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.173444 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdrc2\" (UniqueName: \"kubernetes.io/projected/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-kube-api-access-tdrc2\") pod \"must-gather-mshtd\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.173531 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-must-gather-output\") pod \"must-gather-mshtd\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.274532 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdrc2\" (UniqueName: \"kubernetes.io/projected/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-kube-api-access-tdrc2\") pod \"must-gather-mshtd\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.274606 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-must-gather-output\") pod \"must-gather-mshtd\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.275162 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-must-gather-output\") pod \"must-gather-mshtd\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.291936 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdrc2\" (UniqueName: \"kubernetes.io/projected/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-kube-api-access-tdrc2\") pod \"must-gather-mshtd\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.435131 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 11:58:06 crc kubenswrapper[4758]: I1004 11:58:06.971233 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hm8tb/must-gather-mshtd"] Oct 04 11:58:07 crc kubenswrapper[4758]: I1004 11:58:07.012705 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/must-gather-mshtd" event={"ID":"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835","Type":"ContainerStarted","Data":"10afc148568ae0c2ceae676f778ecac2838d70a15df7f90a2c370597dd2750b8"} Oct 04 11:58:12 crc kubenswrapper[4758]: I1004 11:58:12.056819 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/must-gather-mshtd" event={"ID":"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835","Type":"ContainerStarted","Data":"c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c"} Oct 04 11:58:12 crc kubenswrapper[4758]: I1004 11:58:12.057297 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/must-gather-mshtd" event={"ID":"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835","Type":"ContainerStarted","Data":"51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3"} Oct 04 11:58:12 crc kubenswrapper[4758]: I1004 11:58:12.076446 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hm8tb/must-gather-mshtd" podStartSLOduration=2.177089599 podStartE2EDuration="6.076429258s" podCreationTimestamp="2025-10-04 11:58:06 +0000 UTC" firstStartedPulling="2025-10-04 11:58:06.977704466 +0000 UTC m=+4064.270355355" lastFinishedPulling="2025-10-04 11:58:10.877044125 +0000 UTC m=+4068.169695014" observedRunningTime="2025-10-04 11:58:12.068230518 +0000 UTC m=+4069.360881407" watchObservedRunningTime="2025-10-04 11:58:12.076429258 +0000 UTC m=+4069.369080147" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.067689 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-lg7v2"] Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.069206 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.183215 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc6hn\" (UniqueName: \"kubernetes.io/projected/c8e2f76c-89ea-4d13-85ae-e3150de7b632-kube-api-access-wc6hn\") pod \"crc-debug-lg7v2\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.183353 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c8e2f76c-89ea-4d13-85ae-e3150de7b632-host\") pod \"crc-debug-lg7v2\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.284558 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c8e2f76c-89ea-4d13-85ae-e3150de7b632-host\") pod \"crc-debug-lg7v2\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.284709 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc6hn\" (UniqueName: \"kubernetes.io/projected/c8e2f76c-89ea-4d13-85ae-e3150de7b632-kube-api-access-wc6hn\") pod \"crc-debug-lg7v2\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.284717 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c8e2f76c-89ea-4d13-85ae-e3150de7b632-host\") pod \"crc-debug-lg7v2\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.308659 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc6hn\" (UniqueName: \"kubernetes.io/projected/c8e2f76c-89ea-4d13-85ae-e3150de7b632-kube-api-access-wc6hn\") pod \"crc-debug-lg7v2\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:16 crc kubenswrapper[4758]: I1004 11:58:16.390513 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 11:58:17 crc kubenswrapper[4758]: I1004 11:58:17.111507 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" event={"ID":"c8e2f76c-89ea-4d13-85ae-e3150de7b632","Type":"ContainerStarted","Data":"3c0a228e7e3e39e6e634ea8a13162f937b476cf87200512cdac0b1ae7a6f7d72"} Oct 04 11:58:27 crc kubenswrapper[4758]: I1004 11:58:27.196343 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" event={"ID":"c8e2f76c-89ea-4d13-85ae-e3150de7b632","Type":"ContainerStarted","Data":"a4785a52e630ea7261183a08a34d4e51515482573476e16e453eece68b2a2202"} Oct 04 11:58:27 crc kubenswrapper[4758]: I1004 11:58:27.218841 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" podStartSLOduration=1.051692575 podStartE2EDuration="11.218814707s" podCreationTimestamp="2025-10-04 11:58:16 +0000 UTC" firstStartedPulling="2025-10-04 11:58:16.433516522 +0000 UTC m=+4073.726167411" lastFinishedPulling="2025-10-04 11:58:26.600638654 +0000 UTC m=+4083.893289543" observedRunningTime="2025-10-04 11:58:27.208157711 +0000 UTC m=+4084.500808600" watchObservedRunningTime="2025-10-04 11:58:27.218814707 +0000 UTC m=+4084.511465606" Oct 04 11:58:31 crc kubenswrapper[4758]: I1004 11:58:31.249114 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:58:31 crc kubenswrapper[4758]: I1004 11:58:31.249607 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:59:01 crc kubenswrapper[4758]: I1004 11:59:01.250066 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:59:01 crc kubenswrapper[4758]: I1004 11:59:01.251540 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.668299 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rjbcl"] Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.672675 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.691524 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjbcl"] Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.767207 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-catalog-content\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.767463 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfkvf\" (UniqueName: \"kubernetes.io/projected/70f023e3-d454-4d79-9069-e2dc94687e3d-kube-api-access-xfkvf\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.767545 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-utilities\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.869731 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfkvf\" (UniqueName: \"kubernetes.io/projected/70f023e3-d454-4d79-9069-e2dc94687e3d-kube-api-access-xfkvf\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.869776 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-utilities\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.869829 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-catalog-content\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.870269 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-catalog-content\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.870407 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-utilities\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:07 crc kubenswrapper[4758]: I1004 11:59:07.901796 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfkvf\" (UniqueName: \"kubernetes.io/projected/70f023e3-d454-4d79-9069-e2dc94687e3d-kube-api-access-xfkvf\") pod \"community-operators-rjbcl\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:08 crc kubenswrapper[4758]: I1004 11:59:08.006768 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:08 crc kubenswrapper[4758]: I1004 11:59:08.753023 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjbcl"] Oct 04 11:59:09 crc kubenswrapper[4758]: I1004 11:59:09.547072 4758 generic.go:334] "Generic (PLEG): container finished" podID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerID="e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec" exitCode=0 Oct 04 11:59:09 crc kubenswrapper[4758]: I1004 11:59:09.547239 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerDied","Data":"e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec"} Oct 04 11:59:09 crc kubenswrapper[4758]: I1004 11:59:09.547692 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerStarted","Data":"ebf1c67c708e801cc8bc89d84357afbbb459398f5aa7edd9dfe93bfbbdb0fa60"} Oct 04 11:59:10 crc kubenswrapper[4758]: I1004 11:59:10.564722 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerStarted","Data":"71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458"} Oct 04 11:59:12 crc kubenswrapper[4758]: I1004 11:59:12.586255 4758 generic.go:334] "Generic (PLEG): container finished" podID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerID="71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458" exitCode=0 Oct 04 11:59:12 crc kubenswrapper[4758]: I1004 11:59:12.586382 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerDied","Data":"71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458"} Oct 04 11:59:13 crc kubenswrapper[4758]: I1004 11:59:13.599046 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerStarted","Data":"a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a"} Oct 04 11:59:13 crc kubenswrapper[4758]: I1004 11:59:13.623456 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rjbcl" podStartSLOduration=3.121112353 podStartE2EDuration="6.62343952s" podCreationTimestamp="2025-10-04 11:59:07 +0000 UTC" firstStartedPulling="2025-10-04 11:59:09.548949875 +0000 UTC m=+4126.841600764" lastFinishedPulling="2025-10-04 11:59:13.051277042 +0000 UTC m=+4130.343927931" observedRunningTime="2025-10-04 11:59:13.614361936 +0000 UTC m=+4130.907012825" watchObservedRunningTime="2025-10-04 11:59:13.62343952 +0000 UTC m=+4130.916090409" Oct 04 11:59:18 crc kubenswrapper[4758]: I1004 11:59:18.007433 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:18 crc kubenswrapper[4758]: I1004 11:59:18.007973 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:19 crc kubenswrapper[4758]: I1004 11:59:19.060532 4758 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rjbcl" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="registry-server" probeResult="failure" output=< Oct 04 11:59:19 crc kubenswrapper[4758]: timeout: failed to connect service ":50051" within 1s Oct 04 11:59:19 crc kubenswrapper[4758]: > Oct 04 11:59:28 crc kubenswrapper[4758]: I1004 11:59:28.067059 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:28 crc kubenswrapper[4758]: I1004 11:59:28.127605 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:28 crc kubenswrapper[4758]: I1004 11:59:28.301521 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjbcl"] Oct 04 11:59:29 crc kubenswrapper[4758]: I1004 11:59:29.742046 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rjbcl" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="registry-server" containerID="cri-o://a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a" gracePeriod=2 Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.481520 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.653966 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-utilities\") pod \"70f023e3-d454-4d79-9069-e2dc94687e3d\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.654291 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-catalog-content\") pod \"70f023e3-d454-4d79-9069-e2dc94687e3d\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.654468 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfkvf\" (UniqueName: \"kubernetes.io/projected/70f023e3-d454-4d79-9069-e2dc94687e3d-kube-api-access-xfkvf\") pod \"70f023e3-d454-4d79-9069-e2dc94687e3d\" (UID: \"70f023e3-d454-4d79-9069-e2dc94687e3d\") " Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.661603 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-utilities" (OuterVolumeSpecName: "utilities") pod "70f023e3-d454-4d79-9069-e2dc94687e3d" (UID: "70f023e3-d454-4d79-9069-e2dc94687e3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.661816 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f023e3-d454-4d79-9069-e2dc94687e3d-kube-api-access-xfkvf" (OuterVolumeSpecName: "kube-api-access-xfkvf") pod "70f023e3-d454-4d79-9069-e2dc94687e3d" (UID: "70f023e3-d454-4d79-9069-e2dc94687e3d"). InnerVolumeSpecName "kube-api-access-xfkvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.705005 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70f023e3-d454-4d79-9069-e2dc94687e3d" (UID: "70f023e3-d454-4d79-9069-e2dc94687e3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.750607 4758 generic.go:334] "Generic (PLEG): container finished" podID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerID="a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a" exitCode=0 Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.750660 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerDied","Data":"a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a"} Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.750685 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjbcl" event={"ID":"70f023e3-d454-4d79-9069-e2dc94687e3d","Type":"ContainerDied","Data":"ebf1c67c708e801cc8bc89d84357afbbb459398f5aa7edd9dfe93bfbbdb0fa60"} Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.750696 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjbcl" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.750702 4758 scope.go:117] "RemoveContainer" containerID="a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.758426 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.758471 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70f023e3-d454-4d79-9069-e2dc94687e3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.758483 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfkvf\" (UniqueName: \"kubernetes.io/projected/70f023e3-d454-4d79-9069-e2dc94687e3d-kube-api-access-xfkvf\") on node \"crc\" DevicePath \"\"" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.775597 4758 scope.go:117] "RemoveContainer" containerID="71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.805339 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rjbcl"] Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.809411 4758 scope.go:117] "RemoveContainer" containerID="e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.818132 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rjbcl"] Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.866563 4758 scope.go:117] "RemoveContainer" containerID="a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a" Oct 04 11:59:30 crc kubenswrapper[4758]: E1004 11:59:30.867233 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a\": container with ID starting with a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a not found: ID does not exist" containerID="a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.867267 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a"} err="failed to get container status \"a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a\": rpc error: code = NotFound desc = could not find container \"a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a\": container with ID starting with a1e91717ad96690a7ce8a3aac58f1be2f54280f99ca14fb8055d1d9888cac81a not found: ID does not exist" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.867286 4758 scope.go:117] "RemoveContainer" containerID="71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458" Oct 04 11:59:30 crc kubenswrapper[4758]: E1004 11:59:30.867504 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458\": container with ID starting with 71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458 not found: ID does not exist" containerID="71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.867523 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458"} err="failed to get container status \"71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458\": rpc error: code = NotFound desc = could not find container \"71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458\": container with ID starting with 71570a08f066381355e5f6eb9bcfedaf57ab1a78158099ba4e50af3cc2bf0458 not found: ID does not exist" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.867535 4758 scope.go:117] "RemoveContainer" containerID="e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec" Oct 04 11:59:30 crc kubenswrapper[4758]: E1004 11:59:30.867725 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec\": container with ID starting with e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec not found: ID does not exist" containerID="e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec" Oct 04 11:59:30 crc kubenswrapper[4758]: I1004 11:59:30.867742 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec"} err="failed to get container status \"e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec\": rpc error: code = NotFound desc = could not find container \"e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec\": container with ID starting with e294e633a51535a739abb6209abe83bdc95a12f5a479fe614256c9bb74e69dec not found: ID does not exist" Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.249740 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.250015 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.250058 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.250762 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2320c5922cab48a5f665e70e3a26f8604ef267fa9be39ebc5e9e9f7ea77c98b3"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.250822 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://2320c5922cab48a5f665e70e3a26f8604ef267fa9be39ebc5e9e9f7ea77c98b3" gracePeriod=600 Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.338236 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" path="/var/lib/kubelet/pods/70f023e3-d454-4d79-9069-e2dc94687e3d/volumes" Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.763557 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="2320c5922cab48a5f665e70e3a26f8604ef267fa9be39ebc5e9e9f7ea77c98b3" exitCode=0 Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.763703 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"2320c5922cab48a5f665e70e3a26f8604ef267fa9be39ebc5e9e9f7ea77c98b3"} Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.763898 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerStarted","Data":"57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6"} Oct 04 11:59:31 crc kubenswrapper[4758]: I1004 11:59:31.763919 4758 scope.go:117] "RemoveContainer" containerID="83e0bce53d24a0c9d3277a5227ee655a501bf218daad2b5fa17b158127be98d7" Oct 04 11:59:41 crc kubenswrapper[4758]: I1004 11:59:41.617951 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c9d6c44f4-gx842_4df5fcde-8552-4277-896b-e5f47e9b64b1/barbican-api/0.log" Oct 04 11:59:41 crc kubenswrapper[4758]: I1004 11:59:41.639809 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c9d6c44f4-gx842_4df5fcde-8552-4277-896b-e5f47e9b64b1/barbican-api-log/0.log" Oct 04 11:59:41 crc kubenswrapper[4758]: I1004 11:59:41.926396 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5645f9f854-tfsnf_68f7e592-7d70-4c02-a8c5-9b31ffc06c1c/barbican-keystone-listener-log/0.log" Oct 04 11:59:41 crc kubenswrapper[4758]: I1004 11:59:41.929820 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5645f9f854-tfsnf_68f7e592-7d70-4c02-a8c5-9b31ffc06c1c/barbican-keystone-listener/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.106420 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-75574c7dff-p5nq5_84de7aef-2467-44a1-ae92-c22d87035c0d/barbican-worker/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.198998 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-75574c7dff-p5nq5_84de7aef-2467-44a1-ae92-c22d87035c0d/barbican-worker-log/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.493027 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-78t9c_1b2357af-2674-4b82-b4fe-35ae6c2099d4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.594755 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9e168ea3-6a31-4a16-b478-7b25e649810e/ceilometer-central-agent/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.692826 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9e168ea3-6a31-4a16-b478-7b25e649810e/ceilometer-notification-agent/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.760975 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9e168ea3-6a31-4a16-b478-7b25e649810e/proxy-httpd/0.log" Oct 04 11:59:42 crc kubenswrapper[4758]: I1004 11:59:42.874052 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9e168ea3-6a31-4a16-b478-7b25e649810e/sg-core/0.log" Oct 04 11:59:43 crc kubenswrapper[4758]: I1004 11:59:43.097200 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_49fbaf4c-21ec-4246-86b8-1361ae5f3149/cinder-api/0.log" Oct 04 11:59:43 crc kubenswrapper[4758]: I1004 11:59:43.187452 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_49fbaf4c-21ec-4246-86b8-1361ae5f3149/cinder-api-log/0.log" Oct 04 11:59:43 crc kubenswrapper[4758]: I1004 11:59:43.544965 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_05d625cd-c9ec-4b81-a929-e1b9027e8c8e/cinder-scheduler/0.log" Oct 04 11:59:43 crc kubenswrapper[4758]: I1004 11:59:43.600744 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_05d625cd-c9ec-4b81-a929-e1b9027e8c8e/probe/0.log" Oct 04 11:59:43 crc kubenswrapper[4758]: I1004 11:59:43.911750 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9qwtd_5857921a-d32b-4089-98af-e1b14930e1e9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:44 crc kubenswrapper[4758]: I1004 11:59:44.101666 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4xhgt_e9547006-e0e0-4294-a079-7bb90e5e44ae/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:44 crc kubenswrapper[4758]: I1004 11:59:44.230994 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-n9gzv_c9446f9a-a59b-485b-be69-82419c46d6cd/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:44 crc kubenswrapper[4758]: I1004 11:59:44.438161 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54ffdb7d8c-fz8h2_0acd8097-4dd5-4456-b083-97412d714eb6/init/0.log" Oct 04 11:59:44 crc kubenswrapper[4758]: I1004 11:59:44.763951 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54ffdb7d8c-fz8h2_0acd8097-4dd5-4456-b083-97412d714eb6/dnsmasq-dns/0.log" Oct 04 11:59:44 crc kubenswrapper[4758]: I1004 11:59:44.787445 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54ffdb7d8c-fz8h2_0acd8097-4dd5-4456-b083-97412d714eb6/init/0.log" Oct 04 11:59:44 crc kubenswrapper[4758]: I1004 11:59:44.959445 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-slrft_ceea8611-3691-4c29-b4ae-5c1e0d0bb396/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:45 crc kubenswrapper[4758]: I1004 11:59:45.027074 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7359e803-55a1-4586-bd84-5b389daf4167/glance-httpd/0.log" Oct 04 11:59:45 crc kubenswrapper[4758]: I1004 11:59:45.214637 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_7359e803-55a1-4586-bd84-5b389daf4167/glance-log/0.log" Oct 04 11:59:45 crc kubenswrapper[4758]: I1004 11:59:45.484683 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_69979b78-01c0-491b-8ab3-71aed97a8094/glance-log/0.log" Oct 04 11:59:45 crc kubenswrapper[4758]: I1004 11:59:45.547028 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_69979b78-01c0-491b-8ab3-71aed97a8094/glance-httpd/0.log" Oct 04 11:59:45 crc kubenswrapper[4758]: I1004 11:59:45.788276 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c98cb656b-lczgq_61f721a8-00fb-446c-8e98-922827d2fc39/horizon/1.log" Oct 04 11:59:45 crc kubenswrapper[4758]: I1004 11:59:45.926818 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c98cb656b-lczgq_61f721a8-00fb-446c-8e98-922827d2fc39/horizon/0.log" Oct 04 11:59:46 crc kubenswrapper[4758]: I1004 11:59:46.088354 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c98cb656b-lczgq_61f721a8-00fb-446c-8e98-922827d2fc39/horizon-log/0.log" Oct 04 11:59:46 crc kubenswrapper[4758]: I1004 11:59:46.117195 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jhzdp_bf631c48-0a7f-41cf-911c-c34c42b30558/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:46 crc kubenswrapper[4758]: I1004 11:59:46.248196 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-mlz4q_89420680-fecc-433a-a18e-5f2b65b98dfe/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:46 crc kubenswrapper[4758]: I1004 11:59:46.535936 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_bb9965b2-702d-4b0e-b8ef-2590608e2ebe/kube-state-metrics/0.log" Oct 04 11:59:46 crc kubenswrapper[4758]: I1004 11:59:46.771729 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-ccf54cd5-74s2f_6fdd9c0a-1f42-429f-ace0-2d4b7e37ddce/keystone-api/0.log" Oct 04 11:59:46 crc kubenswrapper[4758]: I1004 11:59:46.885546 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m97jl_6ec33929-038a-47ec-9240-42f88c26168e/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:47 crc kubenswrapper[4758]: I1004 11:59:47.500831 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d66f67fd7-kd9f7_f3b292da-54e8-4646-b70a-b2618eb4af4d/neutron-httpd/0.log" Oct 04 11:59:47 crc kubenswrapper[4758]: I1004 11:59:47.783390 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-8lhhj_2bdca8ac-44ee-4fc1-8b08-7f592ca42a4e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:47 crc kubenswrapper[4758]: I1004 11:59:47.930363 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7d66f67fd7-kd9f7_f3b292da-54e8-4646-b70a-b2618eb4af4d/neutron-api/0.log" Oct 04 11:59:48 crc kubenswrapper[4758]: I1004 11:59:48.822731 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4c88fa4e-31fc-4857-9934-e63c956f00f0/nova-cell0-conductor-conductor/0.log" Oct 04 11:59:49 crc kubenswrapper[4758]: I1004 11:59:49.418447 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d57de266-f458-4d3b-817c-0c33edfcd1f6/nova-api-log/0.log" Oct 04 11:59:49 crc kubenswrapper[4758]: I1004 11:59:49.436122 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_1a824dc0-f283-4e65-9b90-82b0d55e478d/nova-cell1-conductor-conductor/0.log" Oct 04 11:59:49 crc kubenswrapper[4758]: I1004 11:59:49.695062 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d57de266-f458-4d3b-817c-0c33edfcd1f6/nova-api-api/0.log" Oct 04 11:59:49 crc kubenswrapper[4758]: I1004 11:59:49.732367 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_69b3b6d4-8c4d-4dd6-8836-eaf8a4a92033/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 11:59:49 crc kubenswrapper[4758]: I1004 11:59:49.984505 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-cw4t7_02009f56-9d0a-404f-8f19-475ac09aad43/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:50 crc kubenswrapper[4758]: I1004 11:59:50.258329 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_124e2bc8-911c-4842-be08-4674929ed4fb/nova-metadata-log/0.log" Oct 04 11:59:50 crc kubenswrapper[4758]: I1004 11:59:50.922257 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3d86d897-f3e3-4eac-94f4-6fe59fc58d20/mysql-bootstrap/0.log" Oct 04 11:59:50 crc kubenswrapper[4758]: I1004 11:59:50.970705 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_0e96b041-d262-4169-8ee6-cfee13404759/nova-scheduler-scheduler/0.log" Oct 04 11:59:51 crc kubenswrapper[4758]: I1004 11:59:51.205234 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3d86d897-f3e3-4eac-94f4-6fe59fc58d20/mysql-bootstrap/0.log" Oct 04 11:59:51 crc kubenswrapper[4758]: I1004 11:59:51.253439 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_3d86d897-f3e3-4eac-94f4-6fe59fc58d20/galera/0.log" Oct 04 11:59:51 crc kubenswrapper[4758]: I1004 11:59:51.564205 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a/mysql-bootstrap/0.log" Oct 04 11:59:51 crc kubenswrapper[4758]: I1004 11:59:51.687678 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_124e2bc8-911c-4842-be08-4674929ed4fb/nova-metadata-metadata/0.log" Oct 04 11:59:51 crc kubenswrapper[4758]: I1004 11:59:51.742752 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a/mysql-bootstrap/0.log" Oct 04 11:59:51 crc kubenswrapper[4758]: I1004 11:59:51.822922 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae3a69ba-e0ab-4b04-b90a-7bfe4c94078a/galera/0.log" Oct 04 11:59:52 crc kubenswrapper[4758]: I1004 11:59:52.030563 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_78723cfd-8fe5-4c7c-9add-90ac24e14f55/openstackclient/0.log" Oct 04 11:59:52 crc kubenswrapper[4758]: I1004 11:59:52.234091 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gpsfr_39bf6bd9-f043-4485-8980-846341f31b4e/ovn-controller/0.log" Oct 04 11:59:52 crc kubenswrapper[4758]: I1004 11:59:52.435743 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7fwv4_e4d252bc-617d-4014-b340-ce7feb1c7062/ovsdb-server-init/0.log" Oct 04 11:59:52 crc kubenswrapper[4758]: I1004 11:59:52.700719 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7fwv4_e4d252bc-617d-4014-b340-ce7feb1c7062/ovs-vswitchd/0.log" Oct 04 11:59:52 crc kubenswrapper[4758]: I1004 11:59:52.727644 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7fwv4_e4d252bc-617d-4014-b340-ce7feb1c7062/ovsdb-server-init/0.log" Oct 04 11:59:52 crc kubenswrapper[4758]: I1004 11:59:52.771166 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7fwv4_e4d252bc-617d-4014-b340-ce7feb1c7062/ovsdb-server/0.log" Oct 04 11:59:53 crc kubenswrapper[4758]: I1004 11:59:53.142275 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-qdvb5_dfb207a2-ea2f-41bd-8cf5-17ae13aca9f3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:53 crc kubenswrapper[4758]: I1004 11:59:53.287590 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-9cb54f9cb-xrmhr_ed03b984-16f3-46dd-a684-f381b6fbe89b/ovn-northd/0.log" Oct 04 11:59:53 crc kubenswrapper[4758]: I1004 11:59:53.438309 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1c45f9ea-ff92-44d1-801f-2f0faf16597d/ovsdbserver-nb/0.log" Oct 04 11:59:53 crc kubenswrapper[4758]: I1004 11:59:53.650092 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8a3affc8-eb3e-4ab1-ab10-07b690442cff/ovsdbserver-sb/0.log" Oct 04 11:59:54 crc kubenswrapper[4758]: I1004 11:59:54.140755 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8488664d88-ghdnd_96e4bc74-27b8-4c61-85c5-d56589b9e4bf/placement-api/0.log" Oct 04 11:59:54 crc kubenswrapper[4758]: I1004 11:59:54.171713 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8488664d88-ghdnd_96e4bc74-27b8-4c61-85c5-d56589b9e4bf/placement-log/0.log" Oct 04 11:59:54 crc kubenswrapper[4758]: I1004 11:59:54.665234 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_dfdc4c39-424b-4488-9f1a-1efe8f212ac2/setup-container/0.log" Oct 04 11:59:54 crc kubenswrapper[4758]: I1004 11:59:54.841768 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_dfdc4c39-424b-4488-9f1a-1efe8f212ac2/setup-container/0.log" Oct 04 11:59:54 crc kubenswrapper[4758]: I1004 11:59:54.939843 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_dfdc4c39-424b-4488-9f1a-1efe8f212ac2/rabbitmq/0.log" Oct 04 11:59:55 crc kubenswrapper[4758]: I1004 11:59:55.086277 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec7b1391-1a82-4e8b-ba1d-131842a3908c/setup-container/0.log" Oct 04 11:59:55 crc kubenswrapper[4758]: I1004 11:59:55.310256 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec7b1391-1a82-4e8b-ba1d-131842a3908c/setup-container/0.log" Oct 04 11:59:55 crc kubenswrapper[4758]: I1004 11:59:55.391306 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ec7b1391-1a82-4e8b-ba1d-131842a3908c/rabbitmq/0.log" Oct 04 11:59:55 crc kubenswrapper[4758]: I1004 11:59:55.740819 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8zqds_ba6042e4-a924-4ddc-b91f-113c05d8ef70/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:55 crc kubenswrapper[4758]: I1004 11:59:55.742633 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-zml5t_b7817b9e-b878-40b6-b1a5-9ed2526c32a5/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:55 crc kubenswrapper[4758]: I1004 11:59:55.990673 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-k2j8n_daefc3f9-3b67-4223-844d-025717e3094d/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:56 crc kubenswrapper[4758]: I1004 11:59:56.418868 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6szf9_bc1d262a-55f1-46de-8584-fe90c7c82ba5/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:56 crc kubenswrapper[4758]: I1004 11:59:56.477458 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-p8jvs_03f5abcc-d8de-4c25-9664-149f1700a7fb/ssh-known-hosts-edpm-deployment/0.log" Oct 04 11:59:56 crc kubenswrapper[4758]: I1004 11:59:56.752091 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-996d997c5-pbjdp_9170c3d7-3b7f-4945-9e95-76c4b9d476f0/proxy-server/0.log" Oct 04 11:59:56 crc kubenswrapper[4758]: I1004 11:59:56.826486 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-996d997c5-pbjdp_9170c3d7-3b7f-4945-9e95-76c4b9d476f0/proxy-httpd/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.092834 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-b9wqh_a127cd92-1013-49e5-b5ec-832428b0b113/swift-ring-rebalance/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.114244 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/account-auditor/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.379694 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/account-reaper/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.397594 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/account-replicator/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.421379 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/account-server/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.565115 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/container-auditor/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.625632 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/container-server/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.713578 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/container-replicator/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.880168 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/container-updater/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.955164 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/object-auditor/0.log" Oct 04 11:59:57 crc kubenswrapper[4758]: I1004 11:59:57.977989 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/object-expirer/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.201891 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/object-replicator/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.251271 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/object-server/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.251898 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/object-updater/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.403797 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/rsync/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.502991 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d3ed3476-2568-4c27-ac70-87d732389beb/swift-recon-cron/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.724339 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2nbqs_acd41b88-adbc-46b6-b2f0-dde72b71dbf5/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:59:58 crc kubenswrapper[4758]: I1004 11:59:58.949014 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_8a041496-5b8c-4d15-961a-d710b5a461fe/tempest-tests-tempest-tests-runner/0.log" Oct 04 11:59:59 crc kubenswrapper[4758]: I1004 11:59:59.162112 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_2c5474d6-89c4-4089-9681-4be34b6964aa/test-operator-logs-container/0.log" Oct 04 11:59:59 crc kubenswrapper[4758]: I1004 11:59:59.374351 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c9488150-f848-482f-8397-5ec145f84af3/memcached/0.log" Oct 04 11:59:59 crc kubenswrapper[4758]: I1004 11:59:59.396719 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-qxqzh_a36b8558-f3a9-4728-bb2f-f8fe9e6ddbe1/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.140662 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht"] Oct 04 12:00:00 crc kubenswrapper[4758]: E1004 12:00:00.141041 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="extract-content" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.141053 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="extract-content" Oct 04 12:00:00 crc kubenswrapper[4758]: E1004 12:00:00.141077 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="extract-utilities" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.141084 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="extract-utilities" Oct 04 12:00:00 crc kubenswrapper[4758]: E1004 12:00:00.141122 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="registry-server" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.141128 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="registry-server" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.141296 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f023e3-d454-4d79-9069-e2dc94687e3d" containerName="registry-server" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.141872 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.144984 4758 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.153638 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht"] Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.154513 4758 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.205795 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/403336d3-030e-469a-bd3e-13a898f049ea-secret-volume\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.205846 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnxvb\" (UniqueName: \"kubernetes.io/projected/403336d3-030e-469a-bd3e-13a898f049ea-kube-api-access-qnxvb\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.205941 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/403336d3-030e-469a-bd3e-13a898f049ea-config-volume\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.308357 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/403336d3-030e-469a-bd3e-13a898f049ea-secret-volume\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.308782 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnxvb\" (UniqueName: \"kubernetes.io/projected/403336d3-030e-469a-bd3e-13a898f049ea-kube-api-access-qnxvb\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.309035 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/403336d3-030e-469a-bd3e-13a898f049ea-config-volume\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.310383 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/403336d3-030e-469a-bd3e-13a898f049ea-config-volume\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.318389 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/403336d3-030e-469a-bd3e-13a898f049ea-secret-volume\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.327074 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnxvb\" (UniqueName: \"kubernetes.io/projected/403336d3-030e-469a-bd3e-13a898f049ea-kube-api-access-qnxvb\") pod \"collect-profiles-29326320-v9cht\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.462406 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:00 crc kubenswrapper[4758]: I1004 12:00:00.942140 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht"] Oct 04 12:00:01 crc kubenswrapper[4758]: I1004 12:00:01.043704 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" event={"ID":"403336d3-030e-469a-bd3e-13a898f049ea","Type":"ContainerStarted","Data":"b18e643550f8ce83a639269a3a8e92ec4b194b8dccaeef34aae94d171be6eb6c"} Oct 04 12:00:02 crc kubenswrapper[4758]: I1004 12:00:02.069545 4758 generic.go:334] "Generic (PLEG): container finished" podID="403336d3-030e-469a-bd3e-13a898f049ea" containerID="f7ba9fb496b972720672f17bcd91196758d2d9d6b473094c20c47ff512323956" exitCode=0 Oct 04 12:00:02 crc kubenswrapper[4758]: I1004 12:00:02.069721 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" event={"ID":"403336d3-030e-469a-bd3e-13a898f049ea","Type":"ContainerDied","Data":"f7ba9fb496b972720672f17bcd91196758d2d9d6b473094c20c47ff512323956"} Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.792366 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.873998 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/403336d3-030e-469a-bd3e-13a898f049ea-config-volume\") pod \"403336d3-030e-469a-bd3e-13a898f049ea\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.874412 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/403336d3-030e-469a-bd3e-13a898f049ea-secret-volume\") pod \"403336d3-030e-469a-bd3e-13a898f049ea\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.874662 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnxvb\" (UniqueName: \"kubernetes.io/projected/403336d3-030e-469a-bd3e-13a898f049ea-kube-api-access-qnxvb\") pod \"403336d3-030e-469a-bd3e-13a898f049ea\" (UID: \"403336d3-030e-469a-bd3e-13a898f049ea\") " Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.874871 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/403336d3-030e-469a-bd3e-13a898f049ea-config-volume" (OuterVolumeSpecName: "config-volume") pod "403336d3-030e-469a-bd3e-13a898f049ea" (UID: "403336d3-030e-469a-bd3e-13a898f049ea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.875096 4758 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/403336d3-030e-469a-bd3e-13a898f049ea-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.880945 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/403336d3-030e-469a-bd3e-13a898f049ea-kube-api-access-qnxvb" (OuterVolumeSpecName: "kube-api-access-qnxvb") pod "403336d3-030e-469a-bd3e-13a898f049ea" (UID: "403336d3-030e-469a-bd3e-13a898f049ea"). InnerVolumeSpecName "kube-api-access-qnxvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.883927 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/403336d3-030e-469a-bd3e-13a898f049ea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "403336d3-030e-469a-bd3e-13a898f049ea" (UID: "403336d3-030e-469a-bd3e-13a898f049ea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.976347 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnxvb\" (UniqueName: \"kubernetes.io/projected/403336d3-030e-469a-bd3e-13a898f049ea-kube-api-access-qnxvb\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:03 crc kubenswrapper[4758]: I1004 12:00:03.976374 4758 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/403336d3-030e-469a-bd3e-13a898f049ea-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:04 crc kubenswrapper[4758]: I1004 12:00:04.085658 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" event={"ID":"403336d3-030e-469a-bd3e-13a898f049ea","Type":"ContainerDied","Data":"b18e643550f8ce83a639269a3a8e92ec4b194b8dccaeef34aae94d171be6eb6c"} Oct 04 12:00:04 crc kubenswrapper[4758]: I1004 12:00:04.085696 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b18e643550f8ce83a639269a3a8e92ec4b194b8dccaeef34aae94d171be6eb6c" Oct 04 12:00:04 crc kubenswrapper[4758]: I1004 12:00:04.085693 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326320-v9cht" Oct 04 12:00:04 crc kubenswrapper[4758]: I1004 12:00:04.865612 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6"] Oct 04 12:00:04 crc kubenswrapper[4758]: I1004 12:00:04.875507 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-rbpk6"] Oct 04 12:00:05 crc kubenswrapper[4758]: I1004 12:00:05.337444 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="870c503b-cd34-4a52-959c-4b58bbeaf619" path="/var/lib/kubelet/pods/870c503b-cd34-4a52-959c-4b58bbeaf619/volumes" Oct 04 12:00:15 crc kubenswrapper[4758]: I1004 12:00:15.936700 4758 scope.go:117] "RemoveContainer" containerID="723c9b921b1dd264326f3a7d78c5e59b30296488d42b758868322b893dfbeec7" Oct 04 12:00:32 crc kubenswrapper[4758]: I1004 12:00:32.323137 4758 generic.go:334] "Generic (PLEG): container finished" podID="c8e2f76c-89ea-4d13-85ae-e3150de7b632" containerID="a4785a52e630ea7261183a08a34d4e51515482573476e16e453eece68b2a2202" exitCode=0 Oct 04 12:00:32 crc kubenswrapper[4758]: I1004 12:00:32.323328 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" event={"ID":"c8e2f76c-89ea-4d13-85ae-e3150de7b632","Type":"ContainerDied","Data":"a4785a52e630ea7261183a08a34d4e51515482573476e16e453eece68b2a2202"} Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.434555 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.469668 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-lg7v2"] Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.477512 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-lg7v2"] Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.607515 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc6hn\" (UniqueName: \"kubernetes.io/projected/c8e2f76c-89ea-4d13-85ae-e3150de7b632-kube-api-access-wc6hn\") pod \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.607651 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c8e2f76c-89ea-4d13-85ae-e3150de7b632-host\") pod \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\" (UID: \"c8e2f76c-89ea-4d13-85ae-e3150de7b632\") " Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.607787 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8e2f76c-89ea-4d13-85ae-e3150de7b632-host" (OuterVolumeSpecName: "host") pod "c8e2f76c-89ea-4d13-85ae-e3150de7b632" (UID: "c8e2f76c-89ea-4d13-85ae-e3150de7b632"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.608193 4758 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c8e2f76c-89ea-4d13-85ae-e3150de7b632-host\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.625597 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8e2f76c-89ea-4d13-85ae-e3150de7b632-kube-api-access-wc6hn" (OuterVolumeSpecName: "kube-api-access-wc6hn") pod "c8e2f76c-89ea-4d13-85ae-e3150de7b632" (UID: "c8e2f76c-89ea-4d13-85ae-e3150de7b632"). InnerVolumeSpecName "kube-api-access-wc6hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:00:33 crc kubenswrapper[4758]: I1004 12:00:33.709951 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc6hn\" (UniqueName: \"kubernetes.io/projected/c8e2f76c-89ea-4d13-85ae-e3150de7b632-kube-api-access-wc6hn\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.340455 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c0a228e7e3e39e6e634ea8a13162f937b476cf87200512cdac0b1ae7a6f7d72" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.340516 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-lg7v2" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.718807 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-d4rmc"] Oct 04 12:00:34 crc kubenswrapper[4758]: E1004 12:00:34.719709 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="403336d3-030e-469a-bd3e-13a898f049ea" containerName="collect-profiles" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.719730 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="403336d3-030e-469a-bd3e-13a898f049ea" containerName="collect-profiles" Oct 04 12:00:34 crc kubenswrapper[4758]: E1004 12:00:34.719780 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e2f76c-89ea-4d13-85ae-e3150de7b632" containerName="container-00" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.719791 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e2f76c-89ea-4d13-85ae-e3150de7b632" containerName="container-00" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.720095 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="403336d3-030e-469a-bd3e-13a898f049ea" containerName="collect-profiles" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.720229 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8e2f76c-89ea-4d13-85ae-e3150de7b632" containerName="container-00" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.721190 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.833999 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9537fb99-8721-4b9f-abb6-b0b43af71563-host\") pod \"crc-debug-d4rmc\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.834116 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6vpt\" (UniqueName: \"kubernetes.io/projected/9537fb99-8721-4b9f-abb6-b0b43af71563-kube-api-access-w6vpt\") pod \"crc-debug-d4rmc\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.936656 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9537fb99-8721-4b9f-abb6-b0b43af71563-host\") pod \"crc-debug-d4rmc\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.936712 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6vpt\" (UniqueName: \"kubernetes.io/projected/9537fb99-8721-4b9f-abb6-b0b43af71563-kube-api-access-w6vpt\") pod \"crc-debug-d4rmc\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.936801 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9537fb99-8721-4b9f-abb6-b0b43af71563-host\") pod \"crc-debug-d4rmc\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:34 crc kubenswrapper[4758]: I1004 12:00:34.958496 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6vpt\" (UniqueName: \"kubernetes.io/projected/9537fb99-8721-4b9f-abb6-b0b43af71563-kube-api-access-w6vpt\") pod \"crc-debug-d4rmc\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:35 crc kubenswrapper[4758]: I1004 12:00:35.039343 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:35 crc kubenswrapper[4758]: I1004 12:00:35.342122 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8e2f76c-89ea-4d13-85ae-e3150de7b632" path="/var/lib/kubelet/pods/c8e2f76c-89ea-4d13-85ae-e3150de7b632/volumes" Oct 04 12:00:35 crc kubenswrapper[4758]: I1004 12:00:35.348725 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" event={"ID":"9537fb99-8721-4b9f-abb6-b0b43af71563","Type":"ContainerStarted","Data":"7af6d2ac21ee554b4859e8fca0f482d25eb0722c2979107c2e421988274560f5"} Oct 04 12:00:35 crc kubenswrapper[4758]: I1004 12:00:35.348770 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" event={"ID":"9537fb99-8721-4b9f-abb6-b0b43af71563","Type":"ContainerStarted","Data":"343cdd0b812a16dedc2085e957d8f5633a6727599d8f2ffa1d7582401e038fe3"} Oct 04 12:00:35 crc kubenswrapper[4758]: I1004 12:00:35.364046 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" podStartSLOduration=1.364026967 podStartE2EDuration="1.364026967s" podCreationTimestamp="2025-10-04 12:00:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 12:00:35.364017946 +0000 UTC m=+4212.656668835" watchObservedRunningTime="2025-10-04 12:00:35.364026967 +0000 UTC m=+4212.656677856" Oct 04 12:00:36 crc kubenswrapper[4758]: I1004 12:00:36.358737 4758 generic.go:334] "Generic (PLEG): container finished" podID="9537fb99-8721-4b9f-abb6-b0b43af71563" containerID="7af6d2ac21ee554b4859e8fca0f482d25eb0722c2979107c2e421988274560f5" exitCode=0 Oct 04 12:00:36 crc kubenswrapper[4758]: I1004 12:00:36.358781 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" event={"ID":"9537fb99-8721-4b9f-abb6-b0b43af71563","Type":"ContainerDied","Data":"7af6d2ac21ee554b4859e8fca0f482d25eb0722c2979107c2e421988274560f5"} Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.469906 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.584717 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9537fb99-8721-4b9f-abb6-b0b43af71563-host\") pod \"9537fb99-8721-4b9f-abb6-b0b43af71563\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.584893 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6vpt\" (UniqueName: \"kubernetes.io/projected/9537fb99-8721-4b9f-abb6-b0b43af71563-kube-api-access-w6vpt\") pod \"9537fb99-8721-4b9f-abb6-b0b43af71563\" (UID: \"9537fb99-8721-4b9f-abb6-b0b43af71563\") " Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.585243 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9537fb99-8721-4b9f-abb6-b0b43af71563-host" (OuterVolumeSpecName: "host") pod "9537fb99-8721-4b9f-abb6-b0b43af71563" (UID: "9537fb99-8721-4b9f-abb6-b0b43af71563"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.585453 4758 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9537fb99-8721-4b9f-abb6-b0b43af71563-host\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.597698 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9537fb99-8721-4b9f-abb6-b0b43af71563-kube-api-access-w6vpt" (OuterVolumeSpecName: "kube-api-access-w6vpt") pod "9537fb99-8721-4b9f-abb6-b0b43af71563" (UID: "9537fb99-8721-4b9f-abb6-b0b43af71563"). InnerVolumeSpecName "kube-api-access-w6vpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:00:37 crc kubenswrapper[4758]: I1004 12:00:37.686660 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6vpt\" (UniqueName: \"kubernetes.io/projected/9537fb99-8721-4b9f-abb6-b0b43af71563-kube-api-access-w6vpt\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:38 crc kubenswrapper[4758]: I1004 12:00:38.376203 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" event={"ID":"9537fb99-8721-4b9f-abb6-b0b43af71563","Type":"ContainerDied","Data":"343cdd0b812a16dedc2085e957d8f5633a6727599d8f2ffa1d7582401e038fe3"} Oct 04 12:00:38 crc kubenswrapper[4758]: I1004 12:00:38.376532 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="343cdd0b812a16dedc2085e957d8f5633a6727599d8f2ffa1d7582401e038fe3" Oct 04 12:00:38 crc kubenswrapper[4758]: I1004 12:00:38.376595 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-d4rmc" Oct 04 12:00:43 crc kubenswrapper[4758]: I1004 12:00:43.516562 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-d4rmc"] Oct 04 12:00:43 crc kubenswrapper[4758]: I1004 12:00:43.523675 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-d4rmc"] Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.742048 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-wwf94"] Oct 04 12:00:44 crc kubenswrapper[4758]: E1004 12:00:44.742756 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9537fb99-8721-4b9f-abb6-b0b43af71563" containerName="container-00" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.742773 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="9537fb99-8721-4b9f-abb6-b0b43af71563" containerName="container-00" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.742975 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="9537fb99-8721-4b9f-abb6-b0b43af71563" containerName="container-00" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.743653 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.820564 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95fml\" (UniqueName: \"kubernetes.io/projected/48fc8af0-2d46-44b0-8a01-cc396947d979-kube-api-access-95fml\") pod \"crc-debug-wwf94\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.820798 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48fc8af0-2d46-44b0-8a01-cc396947d979-host\") pod \"crc-debug-wwf94\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.923874 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48fc8af0-2d46-44b0-8a01-cc396947d979-host\") pod \"crc-debug-wwf94\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.923966 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95fml\" (UniqueName: \"kubernetes.io/projected/48fc8af0-2d46-44b0-8a01-cc396947d979-kube-api-access-95fml\") pod \"crc-debug-wwf94\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.924049 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48fc8af0-2d46-44b0-8a01-cc396947d979-host\") pod \"crc-debug-wwf94\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:44 crc kubenswrapper[4758]: I1004 12:00:44.952385 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95fml\" (UniqueName: \"kubernetes.io/projected/48fc8af0-2d46-44b0-8a01-cc396947d979-kube-api-access-95fml\") pod \"crc-debug-wwf94\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.059965 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.339891 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9537fb99-8721-4b9f-abb6-b0b43af71563" path="/var/lib/kubelet/pods/9537fb99-8721-4b9f-abb6-b0b43af71563/volumes" Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.446008 4758 generic.go:334] "Generic (PLEG): container finished" podID="48fc8af0-2d46-44b0-8a01-cc396947d979" containerID="94b5d02033e99f9db9670a52aace28ecf2aec0c0fd4af55d78c82e9c0b63f1fb" exitCode=0 Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.446057 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-wwf94" event={"ID":"48fc8af0-2d46-44b0-8a01-cc396947d979","Type":"ContainerDied","Data":"94b5d02033e99f9db9670a52aace28ecf2aec0c0fd4af55d78c82e9c0b63f1fb"} Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.446160 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/crc-debug-wwf94" event={"ID":"48fc8af0-2d46-44b0-8a01-cc396947d979","Type":"ContainerStarted","Data":"cee0c4dd119c6cdf92c0acd2f2680f17a803551143083b359efe9925957b45c7"} Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.493793 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-wwf94"] Oct 04 12:00:45 crc kubenswrapper[4758]: I1004 12:00:45.504907 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hm8tb/crc-debug-wwf94"] Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.550660 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.659270 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95fml\" (UniqueName: \"kubernetes.io/projected/48fc8af0-2d46-44b0-8a01-cc396947d979-kube-api-access-95fml\") pod \"48fc8af0-2d46-44b0-8a01-cc396947d979\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.659460 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48fc8af0-2d46-44b0-8a01-cc396947d979-host\") pod \"48fc8af0-2d46-44b0-8a01-cc396947d979\" (UID: \"48fc8af0-2d46-44b0-8a01-cc396947d979\") " Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.659758 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48fc8af0-2d46-44b0-8a01-cc396947d979-host" (OuterVolumeSpecName: "host") pod "48fc8af0-2d46-44b0-8a01-cc396947d979" (UID: "48fc8af0-2d46-44b0-8a01-cc396947d979"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.660016 4758 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48fc8af0-2d46-44b0-8a01-cc396947d979-host\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.666738 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fc8af0-2d46-44b0-8a01-cc396947d979-kube-api-access-95fml" (OuterVolumeSpecName: "kube-api-access-95fml") pod "48fc8af0-2d46-44b0-8a01-cc396947d979" (UID: "48fc8af0-2d46-44b0-8a01-cc396947d979"). InnerVolumeSpecName "kube-api-access-95fml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:00:46 crc kubenswrapper[4758]: I1004 12:00:46.761280 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95fml\" (UniqueName: \"kubernetes.io/projected/48fc8af0-2d46-44b0-8a01-cc396947d979-kube-api-access-95fml\") on node \"crc\" DevicePath \"\"" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.137159 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d49d7896f-mgphn_6c293821-0ef9-4bf3-9a2a-f188ac3b2151/kube-rbac-proxy/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.220846 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d49d7896f-mgphn_6c293821-0ef9-4bf3-9a2a-f188ac3b2151/manager/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.337011 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fc8af0-2d46-44b0-8a01-cc396947d979" path="/var/lib/kubelet/pods/48fc8af0-2d46-44b0-8a01-cc396947d979/volumes" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.377541 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/util/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.461175 4758 scope.go:117] "RemoveContainer" containerID="94b5d02033e99f9db9670a52aace28ecf2aec0c0fd4af55d78c82e9c0b63f1fb" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.461227 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/crc-debug-wwf94" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.578734 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/util/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.606843 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/pull/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.613001 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/pull/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.762898 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/extract/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.779656 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/pull/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.801452 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bf05bb0378f29be0e82f9f01c89a63b8f46bc7f5aa5e59f3c683a52d93pkl8m_a38fe488-755c-40a2-bec7-9606619aff05/util/0.log" Oct 04 12:00:47 crc kubenswrapper[4758]: I1004 12:00:47.949776 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-866c4585d9-frqpq_112ea09c-f473-4dbc-85d6-1e0ef7e6cf62/kube-rbac-proxy/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.100924 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-866c4585d9-frqpq_112ea09c-f473-4dbc-85d6-1e0ef7e6cf62/manager/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.143566 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-5d8bb764f9-77mnr_59448e96-d587-493a-b5c5-897aa1167152/kube-rbac-proxy/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.288480 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-5d8bb764f9-77mnr_59448e96-d587-493a-b5c5-897aa1167152/manager/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.393057 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6b95d97c7f-lbtmk_e261e6dd-0cdb-43b4-8981-5c34a7288fb3/kube-rbac-proxy/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.407537 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6b95d97c7f-lbtmk_e261e6dd-0cdb-43b4-8981-5c34a7288fb3/manager/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.610344 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-697cc69689-gmq9x_30c739ae-3a42-42c9-babb-03a8dd6b1a50/manager/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.641931 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-697cc69689-gmq9x_30c739ae-3a42-42c9-babb-03a8dd6b1a50/kube-rbac-proxy/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.788036 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-dd56696bb-dm2ls_f22d4cbc-37ae-4621-bc6d-4818642ecd29/kube-rbac-proxy/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.860071 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-dd56696bb-dm2ls_f22d4cbc-37ae-4621-bc6d-4818642ecd29/manager/0.log" Oct 04 12:00:48 crc kubenswrapper[4758]: I1004 12:00:48.873880 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-95c44c5bb-kg2lj_2bafb624-fe21-45cb-b787-192d4353378c/kube-rbac-proxy/0.log" Oct 04 12:00:49 crc kubenswrapper[4758]: I1004 12:00:49.078602 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-578455559c-l7m4w_bfb96416-cb87-48e2-9ea7-71a899c3a620/kube-rbac-proxy/0.log" Oct 04 12:00:49 crc kubenswrapper[4758]: I1004 12:00:49.124249 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-95c44c5bb-kg2lj_2bafb624-fe21-45cb-b787-192d4353378c/manager/0.log" Oct 04 12:00:49 crc kubenswrapper[4758]: I1004 12:00:49.148837 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-578455559c-l7m4w_bfb96416-cb87-48e2-9ea7-71a899c3a620/manager/0.log" Oct 04 12:00:49 crc kubenswrapper[4758]: I1004 12:00:49.921182 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-795ccf785c-jtcj2_aa921475-1667-4eff-8553-9a38928cf00f/kube-rbac-proxy/0.log" Oct 04 12:00:49 crc kubenswrapper[4758]: I1004 12:00:49.960603 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-795ccf785c-jtcj2_aa921475-1667-4eff-8553-9a38928cf00f/manager/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.106728 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6679c59c97-r7gpj_95aa14ab-a652-4eb4-97c1-7ab47431913b/manager/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.156174 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6679c59c97-r7gpj_95aa14ab-a652-4eb4-97c1-7ab47431913b/kube-rbac-proxy/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.178264 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-746ff9ccdc-d269t_7950908c-ddc0-434f-b28a-8d8687392a53/kube-rbac-proxy/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.282001 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-746ff9ccdc-d269t_7950908c-ddc0-434f-b28a-8d8687392a53/manager/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.375647 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-74d4d6c8f6-7ndwj_1f16ce8c-ba77-4c4e-8f48-ee981b138a01/kube-rbac-proxy/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.468086 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-74d4d6c8f6-7ndwj_1f16ce8c-ba77-4c4e-8f48-ee981b138a01/manager/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.644606 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57b88b6ccd-m24qr_66ed67ad-7738-4701-9e08-0e45b5ac73d2/manager/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.648088 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57b88b6ccd-m24qr_66ed67ad-7738-4701-9e08-0e45b5ac73d2/kube-rbac-proxy/0.log" Oct 04 12:00:50 crc kubenswrapper[4758]: I1004 12:00:50.739336 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-768ddcd459-7zwcs_4356f9f3-4e48-4ca8-92d4-f55914ded96b/kube-rbac-proxy/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.182912 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-647c7ff67btjqdt_5b89709e-0a88-4e46-86e6-09e5b1b604da/kube-rbac-proxy/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.184510 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-768ddcd459-7zwcs_4356f9f3-4e48-4ca8-92d4-f55914ded96b/manager/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.211898 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-647c7ff67btjqdt_5b89709e-0a88-4e46-86e6-09e5b1b604da/manager/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.455266 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84d79b6cb5-v9k9r_92f9ca3d-6f51-4fcb-917d-dea0d8bea060/kube-rbac-proxy/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.455696 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-595579fdfd-9p69p_93fd4e69-37c4-485a-b780-6e366ebd0f65/kube-rbac-proxy/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.741711 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84d79b6cb5-v9k9r_92f9ca3d-6f51-4fcb-917d-dea0d8bea060/operator/0.log" Oct 04 12:00:51 crc kubenswrapper[4758]: I1004 12:00:51.765904 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-t92cq_6d7e56fa-428b-4d2f-8378-29806b3e66b4/registry-server/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.039677 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d879db64d-r49wx_cc6e56f4-7374-4ffb-b5f2-55525602b161/kube-rbac-proxy/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.092221 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d879db64d-r49wx_cc6e56f4-7374-4ffb-b5f2-55525602b161/manager/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.117128 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b97cccc57-g8m77_dd449335-3579-4bea-a408-e8b92538f3fa/kube-rbac-proxy/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.316634 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b97cccc57-g8m77_dd449335-3579-4bea-a408-e8b92538f3fa/manager/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.378159 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-595579fdfd-9p69p_93fd4e69-37c4-485a-b780-6e366ebd0f65/manager/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.386407 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-79d8469568-zxtqn_e355cfb0-5886-4e4a-a1cd-48b6e5b442f6/operator/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.542446 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7db4cd47d4-gmj7n_bb6211ba-a514-4243-9204-0eb8954431cd/kube-rbac-proxy/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.578751 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7d9bd964-gqwrj_d9e18e05-fb58-4fa9-b4be-be28657d52ee/kube-rbac-proxy/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.616618 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7db4cd47d4-gmj7n_bb6211ba-a514-4243-9204-0eb8954431cd/manager/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.712099 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5b7d9bd964-gqwrj_d9e18e05-fb58-4fa9-b4be-be28657d52ee/manager/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.809401 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5f4f4847c9-gpsls_5d2ba7ed-1581-4cc2-957d-42ee5bf33781/kube-rbac-proxy/0.log" Oct 04 12:00:52 crc kubenswrapper[4758]: I1004 12:00:52.885414 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5f4f4847c9-gpsls_5d2ba7ed-1581-4cc2-957d-42ee5bf33781/manager/0.log" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.156280 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29326321-gxvdj"] Oct 04 12:01:00 crc kubenswrapper[4758]: E1004 12:01:00.157322 4758 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fc8af0-2d46-44b0-8a01-cc396947d979" containerName="container-00" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.157340 4758 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fc8af0-2d46-44b0-8a01-cc396947d979" containerName="container-00" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.157525 4758 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fc8af0-2d46-44b0-8a01-cc396947d979" containerName="container-00" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.158172 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.177736 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326321-gxvdj"] Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.293610 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-combined-ca-bundle\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.293670 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p42p8\" (UniqueName: \"kubernetes.io/projected/86c249be-8ac2-43ce-8c23-2a940ac311ef-kube-api-access-p42p8\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.293730 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-fernet-keys\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.293755 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-config-data\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.395162 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-combined-ca-bundle\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.395228 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p42p8\" (UniqueName: \"kubernetes.io/projected/86c249be-8ac2-43ce-8c23-2a940ac311ef-kube-api-access-p42p8\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.395276 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-fernet-keys\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.395291 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-config-data\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.406994 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-combined-ca-bundle\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.413433 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-fernet-keys\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.420245 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-config-data\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.420421 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p42p8\" (UniqueName: \"kubernetes.io/projected/86c249be-8ac2-43ce-8c23-2a940ac311ef-kube-api-access-p42p8\") pod \"keystone-cron-29326321-gxvdj\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.478322 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:00 crc kubenswrapper[4758]: I1004 12:01:00.974894 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326321-gxvdj"] Oct 04 12:01:01 crc kubenswrapper[4758]: I1004 12:01:01.575656 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326321-gxvdj" event={"ID":"86c249be-8ac2-43ce-8c23-2a940ac311ef","Type":"ContainerStarted","Data":"c3521ce975ad6c7a2c7bd43ffb278e505a2132e5168176b61381cc242dc94494"} Oct 04 12:01:01 crc kubenswrapper[4758]: I1004 12:01:01.575959 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326321-gxvdj" event={"ID":"86c249be-8ac2-43ce-8c23-2a940ac311ef","Type":"ContainerStarted","Data":"a62ad3232276983a3986c78c5b1c4b76aa936e5ec9f9ebbf87df91622b6422f3"} Oct 04 12:01:01 crc kubenswrapper[4758]: I1004 12:01:01.597517 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29326321-gxvdj" podStartSLOduration=1.5974990949999999 podStartE2EDuration="1.597499095s" podCreationTimestamp="2025-10-04 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 12:01:01.589856179 +0000 UTC m=+4238.882507068" watchObservedRunningTime="2025-10-04 12:01:01.597499095 +0000 UTC m=+4238.890149984" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.128536 4758 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jls5l"] Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.132326 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.146558 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jls5l"] Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.209555 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-catalog-content\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.209606 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-utilities\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.209677 4758 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66kh7\" (UniqueName: \"kubernetes.io/projected/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-kube-api-access-66kh7\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.310910 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-catalog-content\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.310973 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-utilities\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.311062 4758 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66kh7\" (UniqueName: \"kubernetes.io/projected/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-kube-api-access-66kh7\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.311480 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-catalog-content\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.311806 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-utilities\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.349389 4758 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66kh7\" (UniqueName: \"kubernetes.io/projected/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-kube-api-access-66kh7\") pod \"redhat-marketplace-jls5l\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.460012 4758 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.635642 4758 generic.go:334] "Generic (PLEG): container finished" podID="86c249be-8ac2-43ce-8c23-2a940ac311ef" containerID="c3521ce975ad6c7a2c7bd43ffb278e505a2132e5168176b61381cc242dc94494" exitCode=0 Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.635677 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326321-gxvdj" event={"ID":"86c249be-8ac2-43ce-8c23-2a940ac311ef","Type":"ContainerDied","Data":"c3521ce975ad6c7a2c7bd43ffb278e505a2132e5168176b61381cc242dc94494"} Oct 04 12:01:05 crc kubenswrapper[4758]: I1004 12:01:05.962287 4758 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jls5l"] Oct 04 12:01:06 crc kubenswrapper[4758]: I1004 12:01:06.645528 4758 generic.go:334] "Generic (PLEG): container finished" podID="bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" containerID="a62affa45e142546f9ed5bf08534d4f44b167449ef0bb44877c91d2a88fbb1ee" exitCode=0 Oct 04 12:01:06 crc kubenswrapper[4758]: I1004 12:01:06.645588 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jls5l" event={"ID":"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91","Type":"ContainerDied","Data":"a62affa45e142546f9ed5bf08534d4f44b167449ef0bb44877c91d2a88fbb1ee"} Oct 04 12:01:06 crc kubenswrapper[4758]: I1004 12:01:06.645851 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jls5l" event={"ID":"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91","Type":"ContainerStarted","Data":"465a09669c457d6a4777704f74f596c6b7e0d432e2ddc62d848ea9f6bc2af36f"} Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.417335 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.457326 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p42p8\" (UniqueName: \"kubernetes.io/projected/86c249be-8ac2-43ce-8c23-2a940ac311ef-kube-api-access-p42p8\") pod \"86c249be-8ac2-43ce-8c23-2a940ac311ef\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.457495 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-config-data\") pod \"86c249be-8ac2-43ce-8c23-2a940ac311ef\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.457527 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-combined-ca-bundle\") pod \"86c249be-8ac2-43ce-8c23-2a940ac311ef\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.457634 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-fernet-keys\") pod \"86c249be-8ac2-43ce-8c23-2a940ac311ef\" (UID: \"86c249be-8ac2-43ce-8c23-2a940ac311ef\") " Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.475389 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "86c249be-8ac2-43ce-8c23-2a940ac311ef" (UID: "86c249be-8ac2-43ce-8c23-2a940ac311ef"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.478445 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86c249be-8ac2-43ce-8c23-2a940ac311ef-kube-api-access-p42p8" (OuterVolumeSpecName: "kube-api-access-p42p8") pod "86c249be-8ac2-43ce-8c23-2a940ac311ef" (UID: "86c249be-8ac2-43ce-8c23-2a940ac311ef"). InnerVolumeSpecName "kube-api-access-p42p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.550270 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86c249be-8ac2-43ce-8c23-2a940ac311ef" (UID: "86c249be-8ac2-43ce-8c23-2a940ac311ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.560348 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-config-data" (OuterVolumeSpecName: "config-data") pod "86c249be-8ac2-43ce-8c23-2a940ac311ef" (UID: "86c249be-8ac2-43ce-8c23-2a940ac311ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.561234 4758 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.561255 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p42p8\" (UniqueName: \"kubernetes.io/projected/86c249be-8ac2-43ce-8c23-2a940ac311ef-kube-api-access-p42p8\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.561265 4758 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.561274 4758 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86c249be-8ac2-43ce-8c23-2a940ac311ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.654239 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326321-gxvdj" event={"ID":"86c249be-8ac2-43ce-8c23-2a940ac311ef","Type":"ContainerDied","Data":"a62ad3232276983a3986c78c5b1c4b76aa936e5ec9f9ebbf87df91622b6422f3"} Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.654272 4758 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a62ad3232276983a3986c78c5b1c4b76aa936e5ec9f9ebbf87df91622b6422f3" Oct 04 12:01:07 crc kubenswrapper[4758]: I1004 12:01:07.654304 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326321-gxvdj" Oct 04 12:01:08 crc kubenswrapper[4758]: I1004 12:01:08.663250 4758 generic.go:334] "Generic (PLEG): container finished" podID="bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" containerID="21d5b3287e9566cdb91d61585f23b5eb780dc9d4fb1a9acc9bfbdde0b265b7e0" exitCode=0 Oct 04 12:01:08 crc kubenswrapper[4758]: I1004 12:01:08.663533 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jls5l" event={"ID":"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91","Type":"ContainerDied","Data":"21d5b3287e9566cdb91d61585f23b5eb780dc9d4fb1a9acc9bfbdde0b265b7e0"} Oct 04 12:01:09 crc kubenswrapper[4758]: I1004 12:01:09.674626 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jls5l" event={"ID":"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91","Type":"ContainerStarted","Data":"9437dfa900df91888a9d6645cec9349ea611d2f92affdb558dbd76279a32ea6a"} Oct 04 12:01:09 crc kubenswrapper[4758]: I1004 12:01:09.700609 4758 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jls5l" podStartSLOduration=2.253433094 podStartE2EDuration="4.700590981s" podCreationTimestamp="2025-10-04 12:01:05 +0000 UTC" firstStartedPulling="2025-10-04 12:01:06.64751288 +0000 UTC m=+4243.940163769" lastFinishedPulling="2025-10-04 12:01:09.094670767 +0000 UTC m=+4246.387321656" observedRunningTime="2025-10-04 12:01:09.695373281 +0000 UTC m=+4246.988024180" watchObservedRunningTime="2025-10-04 12:01:09.700590981 +0000 UTC m=+4246.993241880" Oct 04 12:01:11 crc kubenswrapper[4758]: I1004 12:01:11.146339 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7mhw4_e6ba4505-f064-4b1d-a545-25e3179485eb/control-plane-machine-set-operator/0.log" Oct 04 12:01:11 crc kubenswrapper[4758]: I1004 12:01:11.206288 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4z96l_62cf7efa-32a9-4564-875e-b7b6b619805f/kube-rbac-proxy/0.log" Oct 04 12:01:11 crc kubenswrapper[4758]: I1004 12:01:11.296056 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4z96l_62cf7efa-32a9-4564-875e-b7b6b619805f/machine-api-operator/0.log" Oct 04 12:01:15 crc kubenswrapper[4758]: I1004 12:01:15.461301 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:15 crc kubenswrapper[4758]: I1004 12:01:15.461916 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:15 crc kubenswrapper[4758]: I1004 12:01:15.515441 4758 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:16 crc kubenswrapper[4758]: I1004 12:01:16.069337 4758 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:16 crc kubenswrapper[4758]: I1004 12:01:16.124178 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jls5l"] Oct 04 12:01:17 crc kubenswrapper[4758]: I1004 12:01:17.741322 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jls5l" podUID="bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" containerName="registry-server" containerID="cri-o://9437dfa900df91888a9d6645cec9349ea611d2f92affdb558dbd76279a32ea6a" gracePeriod=2 Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.753223 4758 generic.go:334] "Generic (PLEG): container finished" podID="bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" containerID="9437dfa900df91888a9d6645cec9349ea611d2f92affdb558dbd76279a32ea6a" exitCode=0 Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.753525 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jls5l" event={"ID":"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91","Type":"ContainerDied","Data":"9437dfa900df91888a9d6645cec9349ea611d2f92affdb558dbd76279a32ea6a"} Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.848909 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.960789 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-utilities\") pod \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.961268 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66kh7\" (UniqueName: \"kubernetes.io/projected/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-kube-api-access-66kh7\") pod \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.961329 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-catalog-content\") pod \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\" (UID: \"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91\") " Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.961958 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-utilities" (OuterVolumeSpecName: "utilities") pod "bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" (UID: "bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.962293 4758 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.972476 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-kube-api-access-66kh7" (OuterVolumeSpecName: "kube-api-access-66kh7") pod "bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" (UID: "bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91"). InnerVolumeSpecName "kube-api-access-66kh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:01:18 crc kubenswrapper[4758]: I1004 12:01:18.983295 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" (UID: "bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.063504 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66kh7\" (UniqueName: \"kubernetes.io/projected/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-kube-api-access-66kh7\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.063792 4758 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.767478 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jls5l" event={"ID":"bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91","Type":"ContainerDied","Data":"465a09669c457d6a4777704f74f596c6b7e0d432e2ddc62d848ea9f6bc2af36f"} Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.768627 4758 scope.go:117] "RemoveContainer" containerID="9437dfa900df91888a9d6645cec9349ea611d2f92affdb558dbd76279a32ea6a" Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.767540 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jls5l" Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.794398 4758 scope.go:117] "RemoveContainer" containerID="21d5b3287e9566cdb91d61585f23b5eb780dc9d4fb1a9acc9bfbdde0b265b7e0" Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.794523 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jls5l"] Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.823024 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jls5l"] Oct 04 12:01:19 crc kubenswrapper[4758]: I1004 12:01:19.829332 4758 scope.go:117] "RemoveContainer" containerID="a62affa45e142546f9ed5bf08534d4f44b167449ef0bb44877c91d2a88fbb1ee" Oct 04 12:01:21 crc kubenswrapper[4758]: I1004 12:01:21.335243 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91" path="/var/lib/kubelet/pods/bbbbc821-fd09-40d2-bbd7-2ab3f13c6b91/volumes" Oct 04 12:01:24 crc kubenswrapper[4758]: I1004 12:01:24.578614 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xmgtx_d2e6b6be-3b22-431d-8401-ce9e45bbb7f1/cert-manager-controller/0.log" Oct 04 12:01:24 crc kubenswrapper[4758]: I1004 12:01:24.724000 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qh9h8_8496f61c-f11f-4358-8f0e-eda4eb9e42df/cert-manager-cainjector/0.log" Oct 04 12:01:24 crc kubenswrapper[4758]: I1004 12:01:24.821350 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-fwh97_0396bbf8-fa84-4c24-916a-0c67c99f1029/cert-manager-webhook/0.log" Oct 04 12:01:31 crc kubenswrapper[4758]: I1004 12:01:31.249261 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 12:01:31 crc kubenswrapper[4758]: I1004 12:01:31.249689 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 12:01:37 crc kubenswrapper[4758]: I1004 12:01:37.411493 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-76lpz_5fb02372-f4a6-408a-8c69-ddee8717758e/nmstate-console-plugin/0.log" Oct 04 12:01:37 crc kubenswrapper[4758]: I1004 12:01:37.438427 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vx78k_30dbe6c8-bfc5-4d5c-bad8-225b4468dbeb/nmstate-handler/0.log" Oct 04 12:01:37 crc kubenswrapper[4758]: I1004 12:01:37.618229 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-f6nvf_ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24/nmstate-metrics/0.log" Oct 04 12:01:37 crc kubenswrapper[4758]: I1004 12:01:37.639969 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-f6nvf_ad5892a8-3bbe-43ab-9544-5ca0f5bcbf24/kube-rbac-proxy/0.log" Oct 04 12:01:37 crc kubenswrapper[4758]: I1004 12:01:37.777901 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-4kt8k_e389e62d-20e1-448b-a6ce-a50609f8dfd0/nmstate-operator/0.log" Oct 04 12:01:37 crc kubenswrapper[4758]: I1004 12:01:37.827669 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-lzss8_ce3569b2-1a2d-4d82-8856-69738a2eb8dd/nmstate-webhook/0.log" Oct 04 12:01:51 crc kubenswrapper[4758]: I1004 12:01:51.863722 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cnb79_063f91cf-fe76-4bc6-a070-ccdf43c43f02/kube-rbac-proxy/0.log" Oct 04 12:01:51 crc kubenswrapper[4758]: I1004 12:01:51.959460 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cnb79_063f91cf-fe76-4bc6-a070-ccdf43c43f02/controller/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.119516 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-frr-files/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.350375 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-frr-files/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.357603 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-metrics/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.397906 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-reloader/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.409469 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-reloader/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.528707 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-reloader/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.586173 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-frr-files/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.618230 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-metrics/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.618920 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-metrics/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.778586 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-frr-files/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.795378 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-reloader/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.847939 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/controller/0.log" Oct 04 12:01:52 crc kubenswrapper[4758]: I1004 12:01:52.875798 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/cp-metrics/0.log" Oct 04 12:01:53 crc kubenswrapper[4758]: I1004 12:01:53.050897 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/kube-rbac-proxy/0.log" Oct 04 12:01:53 crc kubenswrapper[4758]: I1004 12:01:53.103975 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/frr-metrics/0.log" Oct 04 12:01:53 crc kubenswrapper[4758]: I1004 12:01:53.123038 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/kube-rbac-proxy-frr/0.log" Oct 04 12:01:53 crc kubenswrapper[4758]: I1004 12:01:53.327792 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/reloader/0.log" Oct 04 12:01:53 crc kubenswrapper[4758]: I1004 12:01:53.410284 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-48dkt_a20453a0-6ddc-44e5-bd0a-9615f3607127/frr-k8s-webhook-server/0.log" Oct 04 12:01:53 crc kubenswrapper[4758]: I1004 12:01:53.628872 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d8b4d9f4c-wngnq_fa4aec47-b441-4630-91a5-f092db76a1c5/manager/0.log" Oct 04 12:01:54 crc kubenswrapper[4758]: I1004 12:01:54.134272 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-ssxwx_6ae4f2ff-8ed4-4ee5-88ef-38db6e93db5e/frr/0.log" Oct 04 12:01:54 crc kubenswrapper[4758]: I1004 12:01:54.263412 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bz5tp_72277c69-780e-4dcf-a26b-e49001cbf6d1/kube-rbac-proxy/0.log" Oct 04 12:01:54 crc kubenswrapper[4758]: I1004 12:01:54.296689 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6b56c7d7-fmn97_fe3ea4e8-b319-41f2-902b-e3b5dcde02d3/webhook-server/0.log" Oct 04 12:01:54 crc kubenswrapper[4758]: I1004 12:01:54.625931 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bz5tp_72277c69-780e-4dcf-a26b-e49001cbf6d1/speaker/0.log" Oct 04 12:02:01 crc kubenswrapper[4758]: I1004 12:02:01.249422 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 12:02:01 crc kubenswrapper[4758]: I1004 12:02:01.250082 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 12:02:06 crc kubenswrapper[4758]: I1004 12:02:06.996085 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/util/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.192850 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/pull/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.232539 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/util/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.244441 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/pull/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.369159 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/util/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.401629 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/pull/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.451223 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d22rgrk_54cccc46-3ea9-493b-b2fe-20e90218f026/extract/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.565436 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/extract-utilities/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.675289 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/extract-content/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.711433 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/extract-utilities/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.733166 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/extract-content/0.log" Oct 04 12:02:07 crc kubenswrapper[4758]: I1004 12:02:07.956831 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/extract-utilities/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.029741 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/extract-content/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.228959 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/extract-utilities/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.439782 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49qtp_e8129db1-3b11-4a0b-b6b0-2e019c7f45ad/registry-server/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.542994 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/extract-utilities/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.543217 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/extract-content/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.602676 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/extract-content/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.781645 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/extract-content/0.log" Oct 04 12:02:08 crc kubenswrapper[4758]: I1004 12:02:08.850559 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/extract-utilities/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.080828 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/util/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.257592 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-s2xx8_7bf770f2-a1f9-4b2e-80a9-8f6bd13b3d75/registry-server/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.336262 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/util/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.363912 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/pull/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.414033 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/pull/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.572451 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/util/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.580680 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/pull/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.584949 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ckt88c_b4165285-28ee-49e4-901c-06c614575ee9/extract/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.793688 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qhg7v_96f085f8-d3c4-4e51-ab9b-ac854a355e58/marketplace-operator/0.log" Oct 04 12:02:09 crc kubenswrapper[4758]: I1004 12:02:09.876787 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/extract-utilities/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.044383 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/extract-utilities/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.081582 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/extract-content/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.096216 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/extract-content/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.231643 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/extract-utilities/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.238603 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/extract-content/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.421891 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7dwq8_ef2a51a8-cd75-4788-93e1-5f05efa5b0b3/registry-server/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.479457 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/extract-utilities/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.660094 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/extract-content/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.681025 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/extract-utilities/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.685183 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/extract-content/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.861835 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/extract-content/0.log" Oct 04 12:02:10 crc kubenswrapper[4758]: I1004 12:02:10.910567 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/extract-utilities/0.log" Oct 04 12:02:11 crc kubenswrapper[4758]: I1004 12:02:11.302933 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kkfkc_deb271af-edfd-4c2f-ae2b-1a4cbeb64f43/registry-server/0.log" Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.248842 4758 patch_prober.go:28] interesting pod/machine-config-daemon-d4w2q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.249361 4758 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.249406 4758 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.250061 4758 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6"} pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.250123 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerName="machine-config-daemon" containerID="cri-o://57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" gracePeriod=600 Oct 04 12:02:31 crc kubenswrapper[4758]: E1004 12:02:31.369692 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.407597 4758 generic.go:334] "Generic (PLEG): container finished" podID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" exitCode=0 Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.407636 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" event={"ID":"b4fd2c8d-474b-48ac-a369-ce63fbf61413","Type":"ContainerDied","Data":"57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6"} Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.407669 4758 scope.go:117] "RemoveContainer" containerID="2320c5922cab48a5f665e70e3a26f8604ef267fa9be39ebc5e9e9f7ea77c98b3" Oct 04 12:02:31 crc kubenswrapper[4758]: I1004 12:02:31.408321 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:02:31 crc kubenswrapper[4758]: E1004 12:02:31.408557 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:02:34 crc kubenswrapper[4758]: E1004 12:02:34.205137 4758 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.32:58790->38.102.83.32:37165: write tcp 38.102.83.32:58790->38.102.83.32:37165: write: broken pipe Oct 04 12:02:40 crc kubenswrapper[4758]: E1004 12:02:40.844842 4758 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.32:42850->38.102.83.32:37165: read tcp 38.102.83.32:42850->38.102.83.32:37165: read: connection reset by peer Oct 04 12:02:46 crc kubenswrapper[4758]: E1004 12:02:46.303217 4758 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.32:42720->38.102.83.32:37165: write tcp 38.102.83.32:42720->38.102.83.32:37165: write: broken pipe Oct 04 12:02:47 crc kubenswrapper[4758]: I1004 12:02:47.326363 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:02:47 crc kubenswrapper[4758]: E1004 12:02:47.327309 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:03:00 crc kubenswrapper[4758]: I1004 12:03:00.326813 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:03:00 crc kubenswrapper[4758]: E1004 12:03:00.328007 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:03:15 crc kubenswrapper[4758]: I1004 12:03:15.328191 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:03:15 crc kubenswrapper[4758]: E1004 12:03:15.328941 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:03:28 crc kubenswrapper[4758]: I1004 12:03:28.326299 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:03:28 crc kubenswrapper[4758]: E1004 12:03:28.327011 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:03:41 crc kubenswrapper[4758]: I1004 12:03:41.325856 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:03:41 crc kubenswrapper[4758]: E1004 12:03:41.327583 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:03:52 crc kubenswrapper[4758]: I1004 12:03:52.328145 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:03:52 crc kubenswrapper[4758]: E1004 12:03:52.329153 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:04:06 crc kubenswrapper[4758]: I1004 12:04:06.326259 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:04:06 crc kubenswrapper[4758]: E1004 12:04:06.329256 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:04:17 crc kubenswrapper[4758]: I1004 12:04:17.327137 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:04:17 crc kubenswrapper[4758]: E1004 12:04:17.329200 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:04:32 crc kubenswrapper[4758]: I1004 12:04:32.330823 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:04:32 crc kubenswrapper[4758]: E1004 12:04:32.331916 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:04:41 crc kubenswrapper[4758]: I1004 12:04:41.656706 4758 generic.go:334] "Generic (PLEG): container finished" podID="a5e17cb0-d53f-4f2c-8ce0-8193f74e0835" containerID="51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3" exitCode=0 Oct 04 12:04:41 crc kubenswrapper[4758]: I1004 12:04:41.657000 4758 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hm8tb/must-gather-mshtd" event={"ID":"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835","Type":"ContainerDied","Data":"51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3"} Oct 04 12:04:41 crc kubenswrapper[4758]: I1004 12:04:41.657822 4758 scope.go:117] "RemoveContainer" containerID="51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3" Oct 04 12:04:42 crc kubenswrapper[4758]: I1004 12:04:42.528981 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hm8tb_must-gather-mshtd_a5e17cb0-d53f-4f2c-8ce0-8193f74e0835/gather/0.log" Oct 04 12:04:47 crc kubenswrapper[4758]: I1004 12:04:47.326258 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:04:47 crc kubenswrapper[4758]: E1004 12:04:47.327318 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.098365 4758 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hm8tb/must-gather-mshtd"] Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.098927 4758 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hm8tb/must-gather-mshtd" podUID="a5e17cb0-d53f-4f2c-8ce0-8193f74e0835" containerName="copy" containerID="cri-o://c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c" gracePeriod=2 Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.108826 4758 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hm8tb/must-gather-mshtd"] Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.580582 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hm8tb_must-gather-mshtd_a5e17cb0-d53f-4f2c-8ce0-8193f74e0835/copy/0.log" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.581582 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.651697 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdrc2\" (UniqueName: \"kubernetes.io/projected/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-kube-api-access-tdrc2\") pod \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.651834 4758 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-must-gather-output\") pod \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\" (UID: \"a5e17cb0-d53f-4f2c-8ce0-8193f74e0835\") " Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.660407 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-kube-api-access-tdrc2" (OuterVolumeSpecName: "kube-api-access-tdrc2") pod "a5e17cb0-d53f-4f2c-8ce0-8193f74e0835" (UID: "a5e17cb0-d53f-4f2c-8ce0-8193f74e0835"). InnerVolumeSpecName "kube-api-access-tdrc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.754794 4758 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdrc2\" (UniqueName: \"kubernetes.io/projected/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-kube-api-access-tdrc2\") on node \"crc\" DevicePath \"\"" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.792228 4758 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hm8tb_must-gather-mshtd_a5e17cb0-d53f-4f2c-8ce0-8193f74e0835/copy/0.log" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.792888 4758 generic.go:334] "Generic (PLEG): container finished" podID="a5e17cb0-d53f-4f2c-8ce0-8193f74e0835" containerID="c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c" exitCode=143 Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.792942 4758 scope.go:117] "RemoveContainer" containerID="c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.793094 4758 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hm8tb/must-gather-mshtd" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.825989 4758 scope.go:117] "RemoveContainer" containerID="51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.859243 4758 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a5e17cb0-d53f-4f2c-8ce0-8193f74e0835" (UID: "a5e17cb0-d53f-4f2c-8ce0-8193f74e0835"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.884807 4758 scope.go:117] "RemoveContainer" containerID="c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c" Oct 04 12:04:51 crc kubenswrapper[4758]: E1004 12:04:51.885465 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c\": container with ID starting with c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c not found: ID does not exist" containerID="c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.885595 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c"} err="failed to get container status \"c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c\": rpc error: code = NotFound desc = could not find container \"c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c\": container with ID starting with c107464852b818831ae09970d4cd030ea36e827478cb053f12ad23ec72202f5c not found: ID does not exist" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.885689 4758 scope.go:117] "RemoveContainer" containerID="51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3" Oct 04 12:04:51 crc kubenswrapper[4758]: E1004 12:04:51.886052 4758 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3\": container with ID starting with 51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3 not found: ID does not exist" containerID="51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.886084 4758 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3"} err="failed to get container status \"51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3\": rpc error: code = NotFound desc = could not find container \"51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3\": container with ID starting with 51a7cf77925ff0e7d81e7866c97c92abd93bb1ca5902e011f504f01fc1da3cf3 not found: ID does not exist" Oct 04 12:04:51 crc kubenswrapper[4758]: I1004 12:04:51.960625 4758 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 12:04:53 crc kubenswrapper[4758]: I1004 12:04:53.336122 4758 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e17cb0-d53f-4f2c-8ce0-8193f74e0835" path="/var/lib/kubelet/pods/a5e17cb0-d53f-4f2c-8ce0-8193f74e0835/volumes" Oct 04 12:04:59 crc kubenswrapper[4758]: I1004 12:04:59.325855 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:04:59 crc kubenswrapper[4758]: E1004 12:04:59.326922 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:05:14 crc kubenswrapper[4758]: I1004 12:05:14.326553 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:05:14 crc kubenswrapper[4758]: E1004 12:05:14.327567 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:05:16 crc kubenswrapper[4758]: I1004 12:05:16.191875 4758 scope.go:117] "RemoveContainer" containerID="a4785a52e630ea7261183a08a34d4e51515482573476e16e453eece68b2a2202" Oct 04 12:05:27 crc kubenswrapper[4758]: I1004 12:05:27.325726 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:05:27 crc kubenswrapper[4758]: E1004 12:05:27.326451 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:05:41 crc kubenswrapper[4758]: I1004 12:05:41.326367 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:05:41 crc kubenswrapper[4758]: E1004 12:05:41.327009 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:05:56 crc kubenswrapper[4758]: I1004 12:05:56.326217 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:05:56 crc kubenswrapper[4758]: E1004 12:05:56.327445 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:06:07 crc kubenswrapper[4758]: I1004 12:06:07.326775 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:06:07 crc kubenswrapper[4758]: E1004 12:06:07.327935 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:06:18 crc kubenswrapper[4758]: I1004 12:06:18.336319 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:06:18 crc kubenswrapper[4758]: E1004 12:06:18.342366 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:06:30 crc kubenswrapper[4758]: I1004 12:06:30.325963 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:06:30 crc kubenswrapper[4758]: E1004 12:06:30.326810 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" Oct 04 12:06:45 crc kubenswrapper[4758]: I1004 12:06:45.326306 4758 scope.go:117] "RemoveContainer" containerID="57941656b725cdd13f7957a68f975345733ea8256d18964c4829460337d6e1a6" Oct 04 12:06:45 crc kubenswrapper[4758]: E1004 12:06:45.327132 4758 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-d4w2q_openshift-machine-config-operator(b4fd2c8d-474b-48ac-a369-ce63fbf61413)\"" pod="openshift-machine-config-operator/machine-config-daemon-d4w2q" podUID="b4fd2c8d-474b-48ac-a369-ce63fbf61413" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070206746024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070206746017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070175330016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070175331015457 5ustar corecore